var/home/core/zuul-output/0000755000175000017500000000000015116500555014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015116510031015462 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004641010115116510021017662 0ustar rootrootDec 11 08:36:50 crc systemd[1]: Starting Kubernetes Kubelet... Dec 11 08:36:50 crc restorecon[4628]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 11 08:36:50 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:36:51 crc restorecon[4628]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:36:51 crc restorecon[4628]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 11 08:36:52 crc kubenswrapper[4629]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 08:36:52 crc kubenswrapper[4629]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 11 08:36:52 crc kubenswrapper[4629]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 08:36:52 crc kubenswrapper[4629]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 08:36:52 crc kubenswrapper[4629]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 11 08:36:52 crc kubenswrapper[4629]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.052756 4629 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056190 4629 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056208 4629 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056213 4629 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056217 4629 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056223 4629 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056228 4629 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056233 4629 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056237 4629 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056242 4629 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056246 4629 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056250 4629 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056254 4629 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056258 4629 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056262 4629 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056265 4629 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056269 4629 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056280 4629 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056283 4629 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056287 4629 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056290 4629 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056294 4629 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056298 4629 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056301 4629 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056305 4629 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056308 4629 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056312 4629 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056315 4629 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056319 4629 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056322 4629 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056325 4629 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056329 4629 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056332 4629 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056336 4629 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056339 4629 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056343 4629 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056347 4629 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056350 4629 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056353 4629 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056357 4629 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056360 4629 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056364 4629 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056368 4629 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056371 4629 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056376 4629 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056379 4629 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056383 4629 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056386 4629 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056390 4629 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056393 4629 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056396 4629 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056401 4629 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056405 4629 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056410 4629 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056414 4629 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056418 4629 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056422 4629 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056426 4629 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056429 4629 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056433 4629 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056436 4629 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056440 4629 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056444 4629 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056447 4629 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056450 4629 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056454 4629 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056457 4629 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056461 4629 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056464 4629 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056468 4629 feature_gate.go:330] unrecognized feature gate: Example Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056471 4629 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.056477 4629 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056711 4629 flags.go:64] FLAG: --address="0.0.0.0" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056722 4629 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056729 4629 flags.go:64] FLAG: --anonymous-auth="true" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056736 4629 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056743 4629 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056747 4629 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056754 4629 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056759 4629 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056763 4629 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056768 4629 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056772 4629 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056776 4629 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056781 4629 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056785 4629 flags.go:64] FLAG: --cgroup-root="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056789 4629 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056793 4629 flags.go:64] FLAG: --client-ca-file="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056797 4629 flags.go:64] FLAG: --cloud-config="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056801 4629 flags.go:64] FLAG: --cloud-provider="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056805 4629 flags.go:64] FLAG: --cluster-dns="[]" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056809 4629 flags.go:64] FLAG: --cluster-domain="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056813 4629 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056818 4629 flags.go:64] FLAG: --config-dir="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056822 4629 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056826 4629 flags.go:64] FLAG: --container-log-max-files="5" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056832 4629 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056836 4629 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056840 4629 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056858 4629 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056863 4629 flags.go:64] FLAG: --contention-profiling="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056867 4629 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056871 4629 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056875 4629 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056879 4629 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056884 4629 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056888 4629 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056893 4629 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056897 4629 flags.go:64] FLAG: --enable-load-reader="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056902 4629 flags.go:64] FLAG: --enable-server="true" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056911 4629 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056916 4629 flags.go:64] FLAG: --event-burst="100" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056921 4629 flags.go:64] FLAG: --event-qps="50" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056925 4629 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056929 4629 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056933 4629 flags.go:64] FLAG: --eviction-hard="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056938 4629 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056942 4629 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056946 4629 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056950 4629 flags.go:64] FLAG: --eviction-soft="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056955 4629 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056958 4629 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056962 4629 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056966 4629 flags.go:64] FLAG: --experimental-mounter-path="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056970 4629 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056975 4629 flags.go:64] FLAG: --fail-swap-on="true" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056979 4629 flags.go:64] FLAG: --feature-gates="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056984 4629 flags.go:64] FLAG: --file-check-frequency="20s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056988 4629 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056992 4629 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.056996 4629 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057000 4629 flags.go:64] FLAG: --healthz-port="10248" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057005 4629 flags.go:64] FLAG: --help="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057009 4629 flags.go:64] FLAG: --hostname-override="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057012 4629 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057016 4629 flags.go:64] FLAG: --http-check-frequency="20s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057021 4629 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057024 4629 flags.go:64] FLAG: --image-credential-provider-config="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057028 4629 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057032 4629 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057037 4629 flags.go:64] FLAG: --image-service-endpoint="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057041 4629 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057045 4629 flags.go:64] FLAG: --kube-api-burst="100" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057050 4629 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057054 4629 flags.go:64] FLAG: --kube-api-qps="50" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057058 4629 flags.go:64] FLAG: --kube-reserved="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057062 4629 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057066 4629 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057071 4629 flags.go:64] FLAG: --kubelet-cgroups="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057075 4629 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057079 4629 flags.go:64] FLAG: --lock-file="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057083 4629 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057087 4629 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057091 4629 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057097 4629 flags.go:64] FLAG: --log-json-split-stream="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057101 4629 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057105 4629 flags.go:64] FLAG: --log-text-split-stream="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057109 4629 flags.go:64] FLAG: --logging-format="text" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057113 4629 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057117 4629 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057121 4629 flags.go:64] FLAG: --manifest-url="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057125 4629 flags.go:64] FLAG: --manifest-url-header="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057131 4629 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057135 4629 flags.go:64] FLAG: --max-open-files="1000000" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057140 4629 flags.go:64] FLAG: --max-pods="110" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057145 4629 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057149 4629 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057153 4629 flags.go:64] FLAG: --memory-manager-policy="None" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057157 4629 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057162 4629 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057166 4629 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057171 4629 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057182 4629 flags.go:64] FLAG: --node-status-max-images="50" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057186 4629 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057190 4629 flags.go:64] FLAG: --oom-score-adj="-999" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057194 4629 flags.go:64] FLAG: --pod-cidr="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057198 4629 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057204 4629 flags.go:64] FLAG: --pod-manifest-path="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057208 4629 flags.go:64] FLAG: --pod-max-pids="-1" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057212 4629 flags.go:64] FLAG: --pods-per-core="0" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057216 4629 flags.go:64] FLAG: --port="10250" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057221 4629 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057225 4629 flags.go:64] FLAG: --provider-id="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057229 4629 flags.go:64] FLAG: --qos-reserved="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057233 4629 flags.go:64] FLAG: --read-only-port="10255" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057237 4629 flags.go:64] FLAG: --register-node="true" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057241 4629 flags.go:64] FLAG: --register-schedulable="true" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057245 4629 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057252 4629 flags.go:64] FLAG: --registry-burst="10" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057256 4629 flags.go:64] FLAG: --registry-qps="5" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057260 4629 flags.go:64] FLAG: --reserved-cpus="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057264 4629 flags.go:64] FLAG: --reserved-memory="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057269 4629 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057273 4629 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057277 4629 flags.go:64] FLAG: --rotate-certificates="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057281 4629 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057285 4629 flags.go:64] FLAG: --runonce="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057289 4629 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057295 4629 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057299 4629 flags.go:64] FLAG: --seccomp-default="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057303 4629 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057306 4629 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057311 4629 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057315 4629 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057319 4629 flags.go:64] FLAG: --storage-driver-password="root" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057323 4629 flags.go:64] FLAG: --storage-driver-secure="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057327 4629 flags.go:64] FLAG: --storage-driver-table="stats" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057331 4629 flags.go:64] FLAG: --storage-driver-user="root" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057335 4629 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057339 4629 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057343 4629 flags.go:64] FLAG: --system-cgroups="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057347 4629 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057353 4629 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057358 4629 flags.go:64] FLAG: --tls-cert-file="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057362 4629 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057367 4629 flags.go:64] FLAG: --tls-min-version="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057372 4629 flags.go:64] FLAG: --tls-private-key-file="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057376 4629 flags.go:64] FLAG: --topology-manager-policy="none" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057380 4629 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057384 4629 flags.go:64] FLAG: --topology-manager-scope="container" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057388 4629 flags.go:64] FLAG: --v="2" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057395 4629 flags.go:64] FLAG: --version="false" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057400 4629 flags.go:64] FLAG: --vmodule="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057406 4629 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057411 4629 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057517 4629 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057522 4629 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057526 4629 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057530 4629 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057534 4629 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057538 4629 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057541 4629 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057545 4629 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057549 4629 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057553 4629 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057556 4629 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057561 4629 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057587 4629 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057591 4629 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057596 4629 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057601 4629 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057605 4629 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057609 4629 feature_gate.go:330] unrecognized feature gate: Example Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057613 4629 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057617 4629 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057622 4629 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057626 4629 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057630 4629 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057634 4629 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057637 4629 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057641 4629 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057644 4629 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057648 4629 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057651 4629 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057655 4629 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057660 4629 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057663 4629 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057667 4629 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057670 4629 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057673 4629 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057677 4629 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057680 4629 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057684 4629 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057687 4629 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057691 4629 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057694 4629 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057698 4629 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057702 4629 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057706 4629 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057738 4629 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057745 4629 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057752 4629 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057757 4629 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057761 4629 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057765 4629 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057769 4629 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057774 4629 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057778 4629 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057783 4629 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057787 4629 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057792 4629 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057795 4629 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057799 4629 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057802 4629 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057806 4629 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057809 4629 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057813 4629 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057817 4629 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057821 4629 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057824 4629 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057828 4629 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057831 4629 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057835 4629 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057838 4629 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057857 4629 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.057861 4629 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.057868 4629 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.065259 4629 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.065297 4629 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065385 4629 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065397 4629 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065407 4629 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065413 4629 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065419 4629 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065424 4629 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065429 4629 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065435 4629 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065444 4629 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065450 4629 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065455 4629 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065459 4629 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065464 4629 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065469 4629 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065474 4629 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065478 4629 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065483 4629 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065487 4629 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065492 4629 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065497 4629 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065501 4629 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065505 4629 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065509 4629 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065514 4629 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065518 4629 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065532 4629 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065539 4629 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065543 4629 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065549 4629 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065554 4629 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065558 4629 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065562 4629 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065566 4629 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065570 4629 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065574 4629 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065578 4629 feature_gate.go:330] unrecognized feature gate: Example Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065583 4629 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065587 4629 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065591 4629 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065595 4629 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065600 4629 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065603 4629 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065607 4629 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065611 4629 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065616 4629 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065620 4629 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065625 4629 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065629 4629 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065635 4629 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065640 4629 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065645 4629 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065649 4629 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065652 4629 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065656 4629 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065660 4629 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065664 4629 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065668 4629 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065671 4629 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065675 4629 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065679 4629 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065682 4629 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065692 4629 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065696 4629 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065700 4629 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065705 4629 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065713 4629 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065719 4629 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065724 4629 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065730 4629 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065736 4629 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065740 4629 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.065749 4629 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065940 4629 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065948 4629 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065974 4629 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065978 4629 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065982 4629 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065985 4629 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065989 4629 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065992 4629 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065996 4629 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.065999 4629 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066003 4629 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066007 4629 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066014 4629 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066021 4629 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066026 4629 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066030 4629 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066044 4629 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066048 4629 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066065 4629 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066074 4629 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066078 4629 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066082 4629 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066087 4629 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066091 4629 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066095 4629 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066107 4629 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066112 4629 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066116 4629 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066120 4629 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066124 4629 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066128 4629 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066133 4629 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066136 4629 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066141 4629 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066145 4629 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066149 4629 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066153 4629 feature_gate.go:330] unrecognized feature gate: Example Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066157 4629 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066161 4629 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066166 4629 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066170 4629 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066174 4629 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066178 4629 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066184 4629 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066191 4629 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066196 4629 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066201 4629 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066205 4629 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066210 4629 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066215 4629 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066220 4629 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066225 4629 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066229 4629 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066233 4629 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066237 4629 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066241 4629 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066245 4629 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066250 4629 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066255 4629 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066260 4629 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066265 4629 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066278 4629 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066284 4629 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066290 4629 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066295 4629 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066300 4629 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066304 4629 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066308 4629 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066311 4629 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066315 4629 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.066319 4629 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.066325 4629 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.066934 4629 server.go:940] "Client rotation is on, will bootstrap in background" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.072937 4629 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.073074 4629 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.073763 4629 server.go:997] "Starting client certificate rotation" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.073801 4629 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.074319 4629 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-11 19:16:11.084751728 +0000 UTC Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.074500 4629 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 754h39m19.010256107s for next certificate rotation Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.081222 4629 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.083028 4629 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.091460 4629 log.go:25] "Validated CRI v1 runtime API" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.108918 4629 log.go:25] "Validated CRI v1 image API" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.110687 4629 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.113241 4629 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-11-08-30-33-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.113274 4629 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.125984 4629 manager.go:217] Machine: {Timestamp:2025-12-11 08:36:52.124861186 +0000 UTC m=+0.190276814 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199476736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0 BootID:4e353408-98d5-48a9-a921-91df032ed1ab Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:8f:66:76 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:8f:66:76 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:d7:6a:20 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:24:77:60 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:98:37:c1 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:a8:4c:bf Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:56:94:49 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:2a:64:29:55:8f:9a Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:46:07:9d:47:a0:1f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199476736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.126595 4629 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.126800 4629 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.130766 4629 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.130967 4629 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.131004 4629 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.131235 4629 topology_manager.go:138] "Creating topology manager with none policy" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.131248 4629 container_manager_linux.go:303] "Creating device plugin manager" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.131480 4629 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.131513 4629 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.131732 4629 state_mem.go:36] "Initialized new in-memory state store" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.132249 4629 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.133576 4629 kubelet.go:418] "Attempting to sync node with API server" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.133601 4629 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.133649 4629 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.133668 4629 kubelet.go:324] "Adding apiserver pod source" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.133685 4629 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.136233 4629 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.136821 4629 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.137873 4629 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.111:6443: connect: connection refused Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.138051 4629 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.111:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.138372 4629 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.111:6443: connect: connection refused Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.138484 4629 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.111:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.138636 4629 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139306 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139332 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139340 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139347 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139361 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139375 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139388 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139407 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139420 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139437 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139451 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139465 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.139866 4629 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.140413 4629 server.go:1280] "Started kubelet" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.140687 4629 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.140720 4629 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.141070 4629 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.111:6443: connect: connection refused Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.141644 4629 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 11 08:36:52 crc systemd[1]: Started Kubernetes Kubelet. Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.142639 4629 server.go:460] "Adding debug handlers to kubelet server" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.142779 4629 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.142831 4629 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.142930 4629 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 06:45:17.804208211 +0000 UTC Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.143272 4629 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 862h8m25.660938305s for next certificate rotation Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.143440 4629 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.143461 4629 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.143470 4629 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.143573 4629 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.145151 4629 factory.go:153] Registering CRI-O factory Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.145176 4629 factory.go:221] Registration of the crio container factory successfully Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.145230 4629 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.145240 4629 factory.go:55] Registering systemd factory Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.145246 4629 factory.go:221] Registration of the systemd container factory successfully Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.145266 4629 factory.go:103] Registering Raw factory Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.145288 4629 manager.go:1196] Started watching for new ooms in manager Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.145287 4629 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" interval="200ms" Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.145503 4629 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.111:6443: connect: connection refused Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.145793 4629 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.111:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.145806 4629 manager.go:319] Starting recovery of all containers Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.149636 4629 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.111:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18801c56961b325f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 08:36:52.140364383 +0000 UTC m=+0.205779991,LastTimestamp:2025-12-11 08:36:52.140364383 +0000 UTC m=+0.205779991,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.159340 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.159554 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.159641 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.159702 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.159756 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.159811 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.159881 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.159948 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160007 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160062 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160117 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160180 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160235 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160301 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160366 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160422 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160480 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160537 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160604 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160660 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160714 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160767 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160868 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.160942 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161007 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161066 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161126 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161188 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161243 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161333 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161395 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161450 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161506 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161560 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161688 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161747 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161803 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161896 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.161959 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162018 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162078 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162134 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162187 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162241 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162298 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162362 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162416 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162470 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162525 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162578 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162643 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162700 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162761 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162818 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.162889 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163041 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163103 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163167 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163224 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163279 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163339 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163396 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163450 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163503 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163562 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163626 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163691 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163764 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163826 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163922 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.163982 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164048 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164110 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164164 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164219 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164274 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164338 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164397 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164453 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164526 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164584 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164642 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164704 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164764 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164818 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164893 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.164952 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165018 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165077 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165133 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165189 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165243 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165302 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165356 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165410 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165464 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165533 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165599 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165662 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165718 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165772 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165826 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165910 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.165973 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166041 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166100 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166156 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166213 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166301 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166361 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166418 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166489 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166548 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166606 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166669 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166725 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166779 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166834 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166908 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.166974 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167033 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167101 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167157 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167227 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167288 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167343 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167411 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167467 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167521 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167581 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167635 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167690 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167744 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167798 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167866 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167933 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.167994 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168047 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168100 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168159 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168231 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168298 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168353 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168405 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168472 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168534 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168596 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168652 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168706 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168763 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168817 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.168924 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.169541 4629 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.169623 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.169681 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.169737 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.169932 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170031 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170092 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170186 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170247 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170307 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170368 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170424 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170477 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170531 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170592 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170649 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170706 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170798 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170876 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.170939 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171005 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171062 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171118 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171171 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171224 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171284 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171340 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171407 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171465 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171573 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171643 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171733 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171823 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171898 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.171968 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172025 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172119 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172182 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172266 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172352 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172439 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172507 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172564 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172617 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172671 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172727 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172792 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172920 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.172987 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.173068 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.173132 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.173202 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.173297 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.173357 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.173486 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.173551 4629 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.173607 4629 reconstruct.go:97] "Volume reconstruction finished" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.173680 4629 reconciler.go:26] "Reconciler: start to sync state" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.179490 4629 manager.go:324] Recovery completed Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.189587 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.193664 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.193769 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.193827 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.194623 4629 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.196448 4629 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.196544 4629 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.196661 4629 state_mem.go:36] "Initialized new in-memory state store" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.197455 4629 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.197569 4629 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.197636 4629 kubelet.go:2335] "Starting kubelet main sync loop" Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.197766 4629 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.199465 4629 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.111:6443: connect: connection refused Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.199524 4629 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.111:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.206479 4629 policy_none.go:49] "None policy: Start" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.207737 4629 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.207780 4629 state_mem.go:35] "Initializing new in-memory state store" Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.243739 4629 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.265077 4629 manager.go:334] "Starting Device Plugin manager" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.265148 4629 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.265165 4629 server.go:79] "Starting device plugin registration server" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.265631 4629 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.265653 4629 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.266053 4629 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.266143 4629 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.266156 4629 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.274478 4629 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.297943 4629 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.298063 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.299149 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.299175 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.299183 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.299301 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.299730 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.299869 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.301153 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.301180 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.301192 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.301378 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.301547 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.301587 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.302026 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.302058 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.302070 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.302372 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.302401 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.302414 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.303027 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.303164 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.303199 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.303904 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.303923 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.303932 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.304418 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.304436 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.304444 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.304446 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.304465 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.304475 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.304578 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.304706 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.304747 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.305504 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.305528 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.305560 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.305581 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.305593 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.305601 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.305697 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.305718 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.306547 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.306586 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.306596 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.346122 4629 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" interval="400ms" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.366078 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.367314 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.367350 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.367362 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.367391 4629 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.368003 4629 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.111:6443: connect: connection refused" node="crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375267 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375320 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375349 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375417 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375512 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375544 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375562 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375583 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375604 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375627 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375665 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375703 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375736 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375761 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.375783 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477423 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477485 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477503 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477518 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477532 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477547 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477562 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477577 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477597 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477658 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477703 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477702 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477737 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477738 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477710 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477765 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477682 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477776 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477798 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477822 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477886 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477894 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477916 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477918 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477948 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477937 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477996 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.478010 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.477963 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.478137 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.569135 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.570354 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.570393 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.570407 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.570435 4629 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.570957 4629 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.111:6443: connect: connection refused" node="crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.633015 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.639139 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.653540 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.671376 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.674080 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-008e602784aac1732846d635dd908d50081d8e45a5257889529bf2894dfd5ecc WatchSource:0}: Error finding container 008e602784aac1732846d635dd908d50081d8e45a5257889529bf2894dfd5ecc: Status 404 returned error can't find the container with id 008e602784aac1732846d635dd908d50081d8e45a5257889529bf2894dfd5ecc Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.675329 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-f45ec0d7577f21b087f6e7750ffe75a5f4bcfc3c2fb173e27803d31babf2ff38 WatchSource:0}: Error finding container f45ec0d7577f21b087f6e7750ffe75a5f4bcfc3c2fb173e27803d31babf2ff38: Status 404 returned error can't find the container with id f45ec0d7577f21b087f6e7750ffe75a5f4bcfc3c2fb173e27803d31babf2ff38 Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.677671 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.691240 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-0d5513cb0dc7635b8489108e956d5ed02615531f1cd47fbb559536405517836a WatchSource:0}: Error finding container 0d5513cb0dc7635b8489108e956d5ed02615531f1cd47fbb559536405517836a: Status 404 returned error can't find the container with id 0d5513cb0dc7635b8489108e956d5ed02615531f1cd47fbb559536405517836a Dec 11 08:36:52 crc kubenswrapper[4629]: W1211 08:36:52.705781 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-a74ceefbff0ea5e53deec977bca2c5b9a982e1cb2756e04a00d2c2045423d380 WatchSource:0}: Error finding container a74ceefbff0ea5e53deec977bca2c5b9a982e1cb2756e04a00d2c2045423d380: Status 404 returned error can't find the container with id a74ceefbff0ea5e53deec977bca2c5b9a982e1cb2756e04a00d2c2045423d380 Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.747192 4629 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" interval="800ms" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.971589 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.973350 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.973399 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.973415 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:52 crc kubenswrapper[4629]: I1211 08:36:52.973471 4629 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:36:52 crc kubenswrapper[4629]: E1211 08:36:52.974375 4629 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.111:6443: connect: connection refused" node="crc" Dec 11 08:36:53 crc kubenswrapper[4629]: W1211 08:36:53.010508 4629 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.111:6443: connect: connection refused Dec 11 08:36:53 crc kubenswrapper[4629]: E1211 08:36:53.010589 4629 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.111:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:36:53 crc kubenswrapper[4629]: W1211 08:36:53.141050 4629 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.111:6443: connect: connection refused Dec 11 08:36:53 crc kubenswrapper[4629]: E1211 08:36:53.141152 4629 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.111:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.141717 4629 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.111:6443: connect: connection refused Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.204756 4629 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09" exitCode=0 Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.204866 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09"} Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.204999 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a74ceefbff0ea5e53deec977bca2c5b9a982e1cb2756e04a00d2c2045423d380"} Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.205100 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.205886 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.205930 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.205943 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.206506 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4"} Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.206529 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0d5513cb0dc7635b8489108e956d5ed02615531f1cd47fbb559536405517836a"} Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.208196 4629 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838" exitCode=0 Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.208242 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838"} Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.208258 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bf3e32d32ddaf8307cedddf946757721bff7c751b0bd1bd02f84267239216227"} Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.208383 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.209624 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.209644 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.209653 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.209663 4629 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43" exitCode=0 Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.209719 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43"} Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.209738 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f45ec0d7577f21b087f6e7750ffe75a5f4bcfc3c2fb173e27803d31babf2ff38"} Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.209825 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.210625 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.210643 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.210652 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.211561 4629 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="763ac579712e110e6989f12f4e7ab5ed267017bae7b82f3a28b0af7a92274bb8" exitCode=0 Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.211587 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"763ac579712e110e6989f12f4e7ab5ed267017bae7b82f3a28b0af7a92274bb8"} Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.211603 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"008e602784aac1732846d635dd908d50081d8e45a5257889529bf2894dfd5ecc"} Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.211639 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.211653 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.212386 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.212403 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.212424 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.212430 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.212431 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.212443 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:53 crc kubenswrapper[4629]: E1211 08:36:53.548928 4629 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" interval="1.6s" Dec 11 08:36:53 crc kubenswrapper[4629]: W1211 08:36:53.691203 4629 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.111:6443: connect: connection refused Dec 11 08:36:53 crc kubenswrapper[4629]: E1211 08:36:53.691299 4629 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.111:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:36:53 crc kubenswrapper[4629]: W1211 08:36:53.702820 4629 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.111:6443: connect: connection refused Dec 11 08:36:53 crc kubenswrapper[4629]: E1211 08:36:53.702905 4629 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.111:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.775417 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.776950 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.776999 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.777008 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:53 crc kubenswrapper[4629]: I1211 08:36:53.777036 4629 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:36:53 crc kubenswrapper[4629]: E1211 08:36:53.777470 4629 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.111:6443: connect: connection refused" node="crc" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.216667 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4"} Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.216726 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d"} Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.216739 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd"} Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.216751 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468"} Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.218070 4629 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce" exitCode=0 Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.218117 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce"} Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.218248 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.220122 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.220158 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.220169 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.223528 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8edcf812544fff27a0105abbaae88e57c68b9823483aa578a21e79ccd342c8af"} Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.223648 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.225597 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.225665 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.225678 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.229438 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba"} Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.229474 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007"} Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.229490 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5"} Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.229596 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.233689 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.233741 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.233772 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.240874 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a"} Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.240930 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9"} Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.240944 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd"} Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.241005 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.241760 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.241795 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:54 crc kubenswrapper[4629]: I1211 08:36:54.241808 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.246467 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa"} Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.246556 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.247610 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.247636 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.247644 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.249475 4629 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4" exitCode=0 Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.249560 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.249937 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4"} Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.250012 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.250437 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.250469 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.250479 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.250555 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.250570 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.250580 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.378053 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.379309 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.379346 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.379356 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:55 crc kubenswrapper[4629]: I1211 08:36:55.379381 4629 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.105938 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.256884 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607"} Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.256944 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd"} Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.256980 4629 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.257003 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.257015 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.257105 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.256980 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57"} Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.257176 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc"} Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.257196 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01"} Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.258371 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.261888 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.261956 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.261982 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.261889 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.262037 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.261888 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.262087 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.262097 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.262062 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.421618 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.882995 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:56 crc kubenswrapper[4629]: I1211 08:36:56.887711 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.040970 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.259325 4629 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.259399 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.259331 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.259331 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.260312 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.260347 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.260357 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.261170 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.261192 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.261200 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.261514 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.261534 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.261542 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:57 crc kubenswrapper[4629]: I1211 08:36:57.272770 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:58 crc kubenswrapper[4629]: I1211 08:36:58.261267 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:58 crc kubenswrapper[4629]: I1211 08:36:58.261267 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:58 crc kubenswrapper[4629]: I1211 08:36:58.262257 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:58 crc kubenswrapper[4629]: I1211 08:36:58.262293 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:58 crc kubenswrapper[4629]: I1211 08:36:58.262306 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:58 crc kubenswrapper[4629]: I1211 08:36:58.262498 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:58 crc kubenswrapper[4629]: I1211 08:36:58.262611 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:58 crc kubenswrapper[4629]: I1211 08:36:58.262631 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:59 crc kubenswrapper[4629]: I1211 08:36:59.080828 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:36:59 crc kubenswrapper[4629]: I1211 08:36:59.081089 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:59 crc kubenswrapper[4629]: I1211 08:36:59.082593 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:59 crc kubenswrapper[4629]: I1211 08:36:59.082635 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:59 crc kubenswrapper[4629]: I1211 08:36:59.082647 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:36:59 crc kubenswrapper[4629]: I1211 08:36:59.106296 4629 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 08:36:59 crc kubenswrapper[4629]: I1211 08:36:59.106395 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:36:59 crc kubenswrapper[4629]: I1211 08:36:59.784081 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:36:59 crc kubenswrapper[4629]: I1211 08:36:59.784346 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:36:59 crc kubenswrapper[4629]: I1211 08:36:59.785714 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:36:59 crc kubenswrapper[4629]: I1211 08:36:59.785768 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:36:59 crc kubenswrapper[4629]: I1211 08:36:59.785780 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:00 crc kubenswrapper[4629]: I1211 08:37:00.534627 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 11 08:37:00 crc kubenswrapper[4629]: I1211 08:37:00.534827 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:37:00 crc kubenswrapper[4629]: I1211 08:37:00.536009 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:00 crc kubenswrapper[4629]: I1211 08:37:00.536038 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:00 crc kubenswrapper[4629]: I1211 08:37:00.536048 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:02 crc kubenswrapper[4629]: E1211 08:37:02.274579 4629 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 08:37:02 crc kubenswrapper[4629]: I1211 08:37:02.917685 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 11 08:37:02 crc kubenswrapper[4629]: I1211 08:37:02.918130 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:37:02 crc kubenswrapper[4629]: I1211 08:37:02.919899 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:02 crc kubenswrapper[4629]: I1211 08:37:02.919934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:02 crc kubenswrapper[4629]: I1211 08:37:02.919944 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:04 crc kubenswrapper[4629]: E1211 08:37:04.129209 4629 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.18801c56961b325f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 08:36:52.140364383 +0000 UTC m=+0.205779991,LastTimestamp:2025-12-11 08:36:52.140364383 +0000 UTC m=+0.205779991,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 08:37:04 crc kubenswrapper[4629]: I1211 08:37:04.145025 4629 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 11 08:37:04 crc kubenswrapper[4629]: I1211 08:37:04.677779 4629 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 11 08:37:04 crc kubenswrapper[4629]: I1211 08:37:04.677884 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 11 08:37:04 crc kubenswrapper[4629]: I1211 08:37:04.683050 4629 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 11 08:37:04 crc kubenswrapper[4629]: I1211 08:37:04.683376 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 11 08:37:06 crc kubenswrapper[4629]: I1211 08:37:06.262759 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:37:06 crc kubenswrapper[4629]: I1211 08:37:06.262929 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:37:06 crc kubenswrapper[4629]: I1211 08:37:06.264046 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:06 crc kubenswrapper[4629]: I1211 08:37:06.264089 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:06 crc kubenswrapper[4629]: I1211 08:37:06.264101 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:06 crc kubenswrapper[4629]: I1211 08:37:06.429288 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:37:06 crc kubenswrapper[4629]: I1211 08:37:06.429492 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:37:06 crc kubenswrapper[4629]: I1211 08:37:06.430540 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:06 crc kubenswrapper[4629]: I1211 08:37:06.430576 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:06 crc kubenswrapper[4629]: I1211 08:37:06.430588 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:06 crc kubenswrapper[4629]: I1211 08:37:06.433905 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:37:07 crc kubenswrapper[4629]: I1211 08:37:07.284322 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:37:07 crc kubenswrapper[4629]: I1211 08:37:07.285063 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:07 crc kubenswrapper[4629]: I1211 08:37:07.285105 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:07 crc kubenswrapper[4629]: I1211 08:37:07.285119 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:09 crc kubenswrapper[4629]: I1211 08:37:09.107143 4629 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 08:37:09 crc kubenswrapper[4629]: I1211 08:37:09.107238 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 11 08:37:09 crc kubenswrapper[4629]: E1211 08:37:09.662434 4629 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 11 08:37:09 crc kubenswrapper[4629]: I1211 08:37:09.664905 4629 trace.go:236] Trace[2066322361]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 08:36:55.888) (total time: 13776ms): Dec 11 08:37:09 crc kubenswrapper[4629]: Trace[2066322361]: ---"Objects listed" error: 13776ms (08:37:09.664) Dec 11 08:37:09 crc kubenswrapper[4629]: Trace[2066322361]: [13.776421078s] [13.776421078s] END Dec 11 08:37:09 crc kubenswrapper[4629]: I1211 08:37:09.664937 4629 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 11 08:37:09 crc kubenswrapper[4629]: I1211 08:37:09.665325 4629 trace.go:236] Trace[685163841]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 08:36:55.439) (total time: 14225ms): Dec 11 08:37:09 crc kubenswrapper[4629]: Trace[685163841]: ---"Objects listed" error: 14225ms (08:37:09.665) Dec 11 08:37:09 crc kubenswrapper[4629]: Trace[685163841]: [14.225835209s] [14.225835209s] END Dec 11 08:37:09 crc kubenswrapper[4629]: I1211 08:37:09.665363 4629 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 11 08:37:09 crc kubenswrapper[4629]: I1211 08:37:09.665412 4629 trace.go:236] Trace[344517712]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 08:36:56.020) (total time: 13644ms): Dec 11 08:37:09 crc kubenswrapper[4629]: Trace[344517712]: ---"Objects listed" error: 13644ms (08:37:09.665) Dec 11 08:37:09 crc kubenswrapper[4629]: Trace[344517712]: [13.64488884s] [13.64488884s] END Dec 11 08:37:09 crc kubenswrapper[4629]: I1211 08:37:09.665431 4629 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 11 08:37:09 crc kubenswrapper[4629]: I1211 08:37:09.665910 4629 trace.go:236] Trace[2129210066]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 08:36:55.322) (total time: 14343ms): Dec 11 08:37:09 crc kubenswrapper[4629]: Trace[2129210066]: ---"Objects listed" error: 14343ms (08:37:09.665) Dec 11 08:37:09 crc kubenswrapper[4629]: Trace[2129210066]: [14.343382072s] [14.343382072s] END Dec 11 08:37:09 crc kubenswrapper[4629]: I1211 08:37:09.665930 4629 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 11 08:37:09 crc kubenswrapper[4629]: E1211 08:37:09.666816 4629 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 11 08:37:09 crc kubenswrapper[4629]: I1211 08:37:09.667177 4629 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.089118 4629 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:60684->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.089228 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:60684->192.168.126.11:17697: read: connection reset by peer" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.089625 4629 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.089649 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.089795 4629 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.089817 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.143341 4629 apiserver.go:52] "Watching apiserver" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.146458 4629 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.146894 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.147294 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.147459 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.147556 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.147652 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.147722 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.147764 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.147671 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.147726 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.147891 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.148426 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.149800 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.150597 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.150861 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.150968 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.151052 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.152427 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.153555 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.154147 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.178331 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.190537 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.204568 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.223367 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.234705 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.242986 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.244665 4629 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.252488 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.261343 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269563 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269600 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269616 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269633 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269653 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269673 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269696 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269717 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269740 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269766 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269786 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269879 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269908 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269934 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269959 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269980 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.269999 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270020 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270042 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270062 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270084 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270104 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270126 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270132 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270151 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270174 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270199 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270198 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270222 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270245 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270269 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270291 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270314 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270342 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270368 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270392 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270413 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270435 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270505 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270523 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270527 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270542 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270529 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270567 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270593 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270622 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270647 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270670 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270692 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270715 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270741 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270764 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270785 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270806 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270825 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270868 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270890 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270913 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270939 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270962 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270984 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271008 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271036 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271058 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271081 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271107 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271131 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271156 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271184 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271211 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271267 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271298 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271323 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271346 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271372 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271405 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271428 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271451 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271473 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271495 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271518 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271537 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271560 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271577 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271592 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271607 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271641 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271667 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271687 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271712 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271739 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271764 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271782 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271797 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271813 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271828 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271864 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271888 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271907 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271922 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271939 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271957 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271975 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271992 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272010 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272026 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272044 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272064 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272080 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272097 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272112 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272132 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272154 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272177 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272200 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272223 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272245 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272268 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272292 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272311 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272336 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272359 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272386 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272410 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272432 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272455 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272478 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272503 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272527 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272548 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272570 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272591 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272612 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272865 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272897 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272922 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272948 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272971 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272996 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273019 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273042 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273064 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273088 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273113 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273138 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273167 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273192 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273308 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273344 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273369 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273390 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273411 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273430 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273494 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273522 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273546 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273573 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273594 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273615 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273635 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273657 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273683 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273708 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273733 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273758 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273780 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273804 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273830 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273872 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273897 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273924 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273952 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273977 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274002 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274029 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274058 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274081 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274107 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274131 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274155 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274177 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274199 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274224 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274251 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274279 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274303 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274329 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274355 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274378 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274404 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274428 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274454 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274480 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274520 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274546 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274572 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274600 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270718 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274648 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270819 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270903 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270895 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270884 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270911 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.270948 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271039 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271053 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271175 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271181 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271191 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271355 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271371 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271488 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.271944 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272073 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272353 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272399 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272655 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272690 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272837 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.272949 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273101 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273210 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273287 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273338 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273431 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273523 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273558 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273601 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273676 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273799 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273880 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.273994 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274498 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274582 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274615 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274880 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274312 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.275014 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.275705 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.275745 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.275877 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.276048 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.276060 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.276119 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.276810 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.276827 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.276837 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.276977 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.277165 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.277204 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.277448 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.277486 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.277559 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.277529 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.277612 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.277663 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.277896 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.277957 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.278007 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.278041 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.278111 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.278138 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.278197 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.278301 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.278388 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.278499 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.278658 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.278684 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.278834 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.278885 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.279014 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.279182 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.279186 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.279624 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.279648 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.279668 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.279733 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.279761 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:37:10.779732318 +0000 UTC m=+18.845148006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.279765 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.279790 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.280076 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.280081 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.280227 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.280242 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.280415 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.280448 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.280495 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.280515 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.280592 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.280612 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.281101 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.281119 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.281169 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.281208 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.281279 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.281858 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.281995 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.282361 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.282391 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.282412 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.282737 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.282912 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.282932 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.282968 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.283038 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.283111 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.283128 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.283387 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.283449 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.283536 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.283600 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.283627 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.283742 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.283953 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.283979 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.284209 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.284287 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.284637 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.285052 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.285127 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.285135 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.285447 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.285595 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.274626 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.285787 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.285957 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.285985 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286060 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286095 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286420 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286458 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286488 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286512 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286536 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286567 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286600 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286629 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286654 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286680 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286705 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286729 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286810 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286828 4629 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286896 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286909 4629 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286926 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286938 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286950 4629 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286964 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286975 4629 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286983 4629 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286992 4629 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287000 4629 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287009 4629 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287018 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287026 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287035 4629 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287043 4629 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287052 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287061 4629 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287069 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287078 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287086 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287095 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287103 4629 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287113 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287121 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287130 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287139 4629 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287148 4629 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287156 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287164 4629 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287174 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287183 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287192 4629 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287228 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287237 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287247 4629 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287255 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287263 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287273 4629 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287282 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287290 4629 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287301 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287310 4629 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287319 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287328 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287338 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287347 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287360 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287370 4629 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287379 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287387 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287395 4629 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287404 4629 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287412 4629 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287421 4629 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287429 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287438 4629 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287446 4629 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287454 4629 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287462 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287471 4629 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287479 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287489 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287498 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287506 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287515 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287523 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287545 4629 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287558 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287567 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287576 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287585 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287593 4629 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287601 4629 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287610 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287620 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287629 4629 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287638 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287647 4629 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287656 4629 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287665 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287674 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287683 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287691 4629 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287701 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287711 4629 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287720 4629 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287729 4629 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287738 4629 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287725 4629 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.288465 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.288921 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.285857 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.285993 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286349 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286421 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286485 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287752 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290027 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290049 4629 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290063 4629 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290078 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290093 4629 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290107 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290122 4629 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290135 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290148 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290160 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290174 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290187 4629 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290199 4629 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290212 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290223 4629 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290235 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290247 4629 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290258 4629 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290270 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290282 4629 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290295 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290307 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290319 4629 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290331 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290343 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290358 4629 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290370 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290384 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290397 4629 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290410 4629 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290422 4629 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290436 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290450 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290462 4629 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290474 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290487 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290502 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290514 4629 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290525 4629 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.289987 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290538 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286508 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286780 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.286804 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287065 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290550 4629 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287085 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287087 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287400 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290902 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.291146 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287585 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.287761 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.288090 4629 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.291381 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.288177 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.288491 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.291444 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:10.791419391 +0000 UTC m=+18.856835089 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.288549 4629 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.291525 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:10.791509413 +0000 UTC m=+18.856925021 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.288889 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.288934 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.289550 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.289559 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.289543 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.289621 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290140 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290367 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290429 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.290650 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.292652 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.293247 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.293405 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.304667 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.304820 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.304965 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.305098 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.305935 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.305972 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.305987 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.305999 4629 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.306234 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.306298 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.306419 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:10.80640036 +0000 UTC m=+18.871816048 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.306560 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.306610 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.307502 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.308206 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.308231 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.308232 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.308250 4629 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.308297 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:10.80828324 +0000 UTC m=+18.873698858 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.308296 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.308317 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.308362 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.308419 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.308500 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.308658 4629 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa" exitCode=255 Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.308697 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa"} Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.308709 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.309231 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.309482 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.321428 4629 scope.go:117] "RemoveContainer" containerID="f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.321906 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.322104 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.322195 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.323456 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.323999 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.322318 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.324391 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.324739 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.325330 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.325605 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.325693 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.329659 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.324513 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.330376 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.330418 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.337334 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.340928 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.346027 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.350643 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.353455 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.360019 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.364240 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.374042 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.383838 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392280 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392395 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392446 4629 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392479 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392493 4629 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392505 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392518 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392602 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392656 4629 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392810 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392830 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392910 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392929 4629 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392946 4629 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392960 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392976 4629 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.392990 4629 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393005 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393019 4629 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393032 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393047 4629 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393061 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393074 4629 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393086 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393107 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393119 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393132 4629 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393143 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393156 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393169 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393181 4629 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393192 4629 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393205 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393218 4629 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393231 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393243 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393255 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393268 4629 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393283 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393297 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393308 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393320 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393334 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393348 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393362 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393376 4629 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393388 4629 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393399 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393457 4629 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393471 4629 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393482 4629 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393495 4629 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393506 4629 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393517 4629 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393527 4629 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393537 4629 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393547 4629 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393561 4629 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393573 4629 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393584 4629 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393595 4629 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393607 4629 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393618 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393629 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393641 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.393652 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.462055 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.467647 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.473714 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:37:10 crc kubenswrapper[4629]: W1211 08:37:10.492771 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-f4d77605856d33c7163b796d5a56958c569e52c9c8317dc6071b216e29e9f515 WatchSource:0}: Error finding container f4d77605856d33c7163b796d5a56958c569e52c9c8317dc6071b216e29e9f515: Status 404 returned error can't find the container with id f4d77605856d33c7163b796d5a56958c569e52c9c8317dc6071b216e29e9f515 Dec 11 08:37:10 crc kubenswrapper[4629]: W1211 08:37:10.496897 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-8d68c621768e7cd35d1b0177c226e958f1ba371caca777d2269e23b54fb403b7 WatchSource:0}: Error finding container 8d68c621768e7cd35d1b0177c226e958f1ba371caca777d2269e23b54fb403b7: Status 404 returned error can't find the container with id 8d68c621768e7cd35d1b0177c226e958f1ba371caca777d2269e23b54fb403b7 Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.823591 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.823681 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.823713 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.823732 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.823806 4629 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.823807 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:37:11.823774731 +0000 UTC m=+19.889190339 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.823905 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:11.823888865 +0000 UTC m=+19.889304473 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: I1211 08:37:10.823920 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.823984 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.824011 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.824016 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.824024 4629 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.824031 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.824043 4629 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.824072 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:11.824064161 +0000 UTC m=+19.889479779 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.824086 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:11.824081001 +0000 UTC m=+19.889496609 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.824114 4629 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:10 crc kubenswrapper[4629]: E1211 08:37:10.824148 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:11.824139343 +0000 UTC m=+19.889555061 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.169104 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-n44pt"] Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.169461 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-xq9dz"] Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.169639 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-n44pt" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.170668 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xq9dz" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.173057 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.173443 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.173635 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.174030 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.174495 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.174516 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.175733 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.195823 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.240473 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.269578 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.309180 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.313195 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.315210 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da"} Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.315586 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.316268 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8d68c621768e7cd35d1b0177c226e958f1ba371caca777d2269e23b54fb403b7"} Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.319082 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea"} Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.319107 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd"} Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.319120 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f4d77605856d33c7163b796d5a56958c569e52c9c8317dc6071b216e29e9f515"} Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.320710 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85"} Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.320767 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5de5ca3c14d91a1d759093942298354364b4e53028930de957c799f0b769abe8"} Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.328102 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e908f308-e6e4-4a2e-ab9f-48e98698f2f7-hosts-file\") pod \"node-resolver-n44pt\" (UID: \"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\") " pod="openshift-dns/node-resolver-n44pt" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.328144 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa83f85c-0050-4297-b65c-4837ecaf91b3-host\") pod \"node-ca-xq9dz\" (UID: \"aa83f85c-0050-4297-b65c-4837ecaf91b3\") " pod="openshift-image-registry/node-ca-xq9dz" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.328193 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57vgn\" (UniqueName: \"kubernetes.io/projected/e908f308-e6e4-4a2e-ab9f-48e98698f2f7-kube-api-access-57vgn\") pod \"node-resolver-n44pt\" (UID: \"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\") " pod="openshift-dns/node-resolver-n44pt" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.328225 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55vjk\" (UniqueName: \"kubernetes.io/projected/aa83f85c-0050-4297-b65c-4837ecaf91b3-kube-api-access-55vjk\") pod \"node-ca-xq9dz\" (UID: \"aa83f85c-0050-4297-b65c-4837ecaf91b3\") " pod="openshift-image-registry/node-ca-xq9dz" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.328278 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/aa83f85c-0050-4297-b65c-4837ecaf91b3-serviceca\") pod \"node-ca-xq9dz\" (UID: \"aa83f85c-0050-4297-b65c-4837ecaf91b3\") " pod="openshift-image-registry/node-ca-xq9dz" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.352756 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.364086 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.383033 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.399132 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.418959 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.428821 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa83f85c-0050-4297-b65c-4837ecaf91b3-host\") pod \"node-ca-xq9dz\" (UID: \"aa83f85c-0050-4297-b65c-4837ecaf91b3\") " pod="openshift-image-registry/node-ca-xq9dz" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.428875 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e908f308-e6e4-4a2e-ab9f-48e98698f2f7-hosts-file\") pod \"node-resolver-n44pt\" (UID: \"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\") " pod="openshift-dns/node-resolver-n44pt" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.428952 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57vgn\" (UniqueName: \"kubernetes.io/projected/e908f308-e6e4-4a2e-ab9f-48e98698f2f7-kube-api-access-57vgn\") pod \"node-resolver-n44pt\" (UID: \"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\") " pod="openshift-dns/node-resolver-n44pt" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.428988 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55vjk\" (UniqueName: \"kubernetes.io/projected/aa83f85c-0050-4297-b65c-4837ecaf91b3-kube-api-access-55vjk\") pod \"node-ca-xq9dz\" (UID: \"aa83f85c-0050-4297-b65c-4837ecaf91b3\") " pod="openshift-image-registry/node-ca-xq9dz" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.429047 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/aa83f85c-0050-4297-b65c-4837ecaf91b3-serviceca\") pod \"node-ca-xq9dz\" (UID: \"aa83f85c-0050-4297-b65c-4837ecaf91b3\") " pod="openshift-image-registry/node-ca-xq9dz" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.430164 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa83f85c-0050-4297-b65c-4837ecaf91b3-host\") pod \"node-ca-xq9dz\" (UID: \"aa83f85c-0050-4297-b65c-4837ecaf91b3\") " pod="openshift-image-registry/node-ca-xq9dz" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.430255 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e908f308-e6e4-4a2e-ab9f-48e98698f2f7-hosts-file\") pod \"node-resolver-n44pt\" (UID: \"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\") " pod="openshift-dns/node-resolver-n44pt" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.433323 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.448314 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.448985 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55vjk\" (UniqueName: \"kubernetes.io/projected/aa83f85c-0050-4297-b65c-4837ecaf91b3-kube-api-access-55vjk\") pod \"node-ca-xq9dz\" (UID: \"aa83f85c-0050-4297-b65c-4837ecaf91b3\") " pod="openshift-image-registry/node-ca-xq9dz" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.452523 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/aa83f85c-0050-4297-b65c-4837ecaf91b3-serviceca\") pod \"node-ca-xq9dz\" (UID: \"aa83f85c-0050-4297-b65c-4837ecaf91b3\") " pod="openshift-image-registry/node-ca-xq9dz" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.455640 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57vgn\" (UniqueName: \"kubernetes.io/projected/e908f308-e6e4-4a2e-ab9f-48e98698f2f7-kube-api-access-57vgn\") pod \"node-resolver-n44pt\" (UID: \"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\") " pod="openshift-dns/node-resolver-n44pt" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.473680 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.484679 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-n44pt" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.492864 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xq9dz" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.505258 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.523801 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.539330 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.561692 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.580674 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.633696 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-wbld9"] Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.634201 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.635122 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-lvjmx"] Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.635311 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: W1211 08:37:11.635878 4629 reflector.go:561] object-"openshift-machine-config-operator"/"proxy-tls": failed to list *v1.Secret: secrets "proxy-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.635934 4629 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"proxy-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"proxy-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 11 08:37:11 crc kubenswrapper[4629]: W1211 08:37:11.636015 4629 reflector.go:561] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": failed to list *v1.Secret: secrets "machine-config-daemon-dockercfg-r5tcq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.636043 4629 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-r5tcq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-config-daemon-dockercfg-r5tcq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 11 08:37:11 crc kubenswrapper[4629]: W1211 08:37:11.636919 4629 reflector.go:561] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.636948 4629 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 11 08:37:11 crc kubenswrapper[4629]: W1211 08:37:11.637071 4629 reflector.go:561] object-"openshift-multus"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.637091 4629 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 11 08:37:11 crc kubenswrapper[4629]: W1211 08:37:11.637129 4629 reflector.go:561] object-"openshift-multus"/"cni-copy-resources": failed to list *v1.ConfigMap: configmaps "cni-copy-resources" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.637141 4629 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"cni-copy-resources\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cni-copy-resources\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 11 08:37:11 crc kubenswrapper[4629]: W1211 08:37:11.637279 4629 reflector.go:561] object-"openshift-machine-config-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.637295 4629 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 11 08:37:11 crc kubenswrapper[4629]: W1211 08:37:11.637339 4629 reflector.go:561] object-"openshift-machine-config-operator"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.637353 4629 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 11 08:37:11 crc kubenswrapper[4629]: W1211 08:37:11.637392 4629 reflector.go:561] object-"openshift-multus"/"multus-daemon-config": failed to list *v1.ConfigMap: configmaps "multus-daemon-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.637404 4629 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-daemon-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"multus-daemon-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 11 08:37:11 crc kubenswrapper[4629]: W1211 08:37:11.637440 4629 reflector.go:561] object-"openshift-multus"/"default-dockercfg-2q5b6": failed to list *v1.Secret: secrets "default-dockercfg-2q5b6" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.637453 4629 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-dockercfg-2q5b6\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-2q5b6\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 11 08:37:11 crc kubenswrapper[4629]: W1211 08:37:11.637608 4629 reflector.go:561] object-"openshift-multus"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.637633 4629 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.639956 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-gn7n4"] Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.640549 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.642129 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.642723 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.648356 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: W1211 08:37:11.675253 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa83f85c_0050_4297_b65c_4837ecaf91b3.slice/crio-3f466dfb0829b69d75b608d3d673c6e0009256f7d01e22bdbccf2bbc64cfe642 WatchSource:0}: Error finding container 3f466dfb0829b69d75b608d3d673c6e0009256f7d01e22bdbccf2bbc64cfe642: Status 404 returned error can't find the container with id 3f466dfb0829b69d75b608d3d673c6e0009256f7d01e22bdbccf2bbc64cfe642 Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.687286 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.731301 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0693c14e-aca1-4ada-8993-b17e90f51798-os-release\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.731576 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-multus-socket-dir-parent\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.731663 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-hostroot\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.731818 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0693c14e-aca1-4ada-8993-b17e90f51798-cnibin\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.731953 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-os-release\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.732061 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0693c14e-aca1-4ada-8993-b17e90f51798-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.732182 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-549hm\" (UniqueName: \"kubernetes.io/projected/847b50b9-0093-404f-b0dd-63c814d8986b-kube-api-access-549hm\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.732280 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0693c14e-aca1-4ada-8993-b17e90f51798-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.732414 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-var-lib-cni-multus\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.732588 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-multus-cni-dir\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.732695 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfwqs\" (UniqueName: \"kubernetes.io/projected/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-kube-api-access-xfwqs\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.732819 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-var-lib-cni-bin\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.732940 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-run-netns\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.733040 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-var-lib-kubelet\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.733147 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-system-cni-dir\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.733244 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0693c14e-aca1-4ada-8993-b17e90f51798-system-cni-dir\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.733356 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-run-multus-certs\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.733450 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/847b50b9-0093-404f-b0dd-63c814d8986b-mcd-auth-proxy-config\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.733544 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-run-k8s-cni-cncf-io\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.733815 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-cni-binary-copy\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.733904 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-etc-kubernetes\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.733987 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/847b50b9-0093-404f-b0dd-63c814d8986b-rootfs\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.734025 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/847b50b9-0093-404f-b0dd-63c814d8986b-proxy-tls\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.734060 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-cnibin\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.734093 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-multus-conf-dir\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.734117 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-multus-daemon-config\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.734142 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqb45\" (UniqueName: \"kubernetes.io/projected/0693c14e-aca1-4ada-8993-b17e90f51798-kube-api-access-tqb45\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.734193 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0693c14e-aca1-4ada-8993-b17e90f51798-cni-binary-copy\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.756213 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.809264 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835321 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835411 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-multus-socket-dir-parent\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835433 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-hostroot\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835510 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-hostroot\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.835543 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:37:13.83550646 +0000 UTC m=+21.900922068 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835604 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0693c14e-aca1-4ada-8993-b17e90f51798-cnibin\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835630 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0693c14e-aca1-4ada-8993-b17e90f51798-cnibin\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835641 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0693c14e-aca1-4ada-8993-b17e90f51798-os-release\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835663 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-os-release\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835679 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-multus-socket-dir-parent\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835685 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0693c14e-aca1-4ada-8993-b17e90f51798-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835736 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-549hm\" (UniqueName: \"kubernetes.io/projected/847b50b9-0093-404f-b0dd-63c814d8986b-kube-api-access-549hm\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835761 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0693c14e-aca1-4ada-8993-b17e90f51798-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835788 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835804 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-os-release\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835857 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-var-lib-cni-multus\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835808 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-var-lib-cni-multus\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.835919 4629 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835929 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835954 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-multus-cni-dir\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.835977 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:13.835959584 +0000 UTC m=+21.901375282 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835995 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfwqs\" (UniqueName: \"kubernetes.io/projected/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-kube-api-access-xfwqs\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836022 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836038 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0693c14e-aca1-4ada-8993-b17e90f51798-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836047 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-var-lib-cni-bin\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836068 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-system-cni-dir\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836089 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-run-netns\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.835830 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0693c14e-aca1-4ada-8993-b17e90f51798-os-release\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836116 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-var-lib-kubelet\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836144 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-var-lib-kubelet\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836166 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-multus-cni-dir\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836169 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0693c14e-aca1-4ada-8993-b17e90f51798-system-cni-dir\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836207 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-run-k8s-cni-cncf-io\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836235 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-run-multus-certs\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836273 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/847b50b9-0093-404f-b0dd-63c814d8986b-mcd-auth-proxy-config\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836298 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-system-cni-dir\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836297 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836333 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-cni-binary-copy\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836350 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-etc-kubernetes\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836366 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-cnibin\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836383 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-multus-conf-dir\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836388 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0693c14e-aca1-4ada-8993-b17e90f51798-system-cni-dir\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836400 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/847b50b9-0093-404f-b0dd-63c814d8986b-rootfs\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836427 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/847b50b9-0093-404f-b0dd-63c814d8986b-proxy-tls\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836451 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-run-multus-certs\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.836215 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836475 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-multus-daemon-config\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.836489 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836494 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqb45\" (UniqueName: \"kubernetes.io/projected/0693c14e-aca1-4ada-8993-b17e90f51798-kube-api-access-tqb45\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.836502 4629 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836512 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0693c14e-aca1-4ada-8993-b17e90f51798-cni-binary-copy\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836236 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-var-lib-cni-bin\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836429 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-etc-kubernetes\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.836536 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:13.836526222 +0000 UTC m=+21.901941930 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.836300 4629 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836575 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-multus-conf-dir\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836586 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-cnibin\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836604 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/847b50b9-0093-404f-b0dd-63c814d8986b-rootfs\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836614 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0693c14e-aca1-4ada-8993-b17e90f51798-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.836609 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:13.836597044 +0000 UTC m=+21.902012722 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836452 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-run-netns\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.836426 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-host-run-k8s-cni-cncf-io\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.836366 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.836668 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.836677 4629 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:11 crc kubenswrapper[4629]: E1211 08:37:11.836701 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:13.836693467 +0000 UTC m=+21.902109075 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.868626 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.908218 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.953075 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.969805 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:11 crc kubenswrapper[4629]: I1211 08:37:11.997382 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.017701 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.037675 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.075154 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.109069 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.134883 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lt499"] Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.144874 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.152182 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.155780 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.156098 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.156345 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.156485 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.158423 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.158466 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.158626 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.198199 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.198271 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.198304 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:12 crc kubenswrapper[4629]: E1211 08:37:12.198354 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:12 crc kubenswrapper[4629]: E1211 08:37:12.198412 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:12 crc kubenswrapper[4629]: E1211 08:37:12.198449 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.202436 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.203275 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.204633 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.205539 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.206381 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.208374 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.209280 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.210206 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.211608 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.212541 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.213816 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.215006 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.217261 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.218077 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.218997 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.220365 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.221500 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.222529 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.225676 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.226707 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.228060 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.228651 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.229429 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.230610 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.231547 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.234402 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.235397 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.237446 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.238339 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.240016 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.240288 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.241353 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.242067 4629 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.242230 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245202 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245645 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-systemd-units\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245700 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-ovn\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245720 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-node-log\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245747 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-kubelet\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245772 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-run-netns\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245792 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f89bfa10-15c0-40cd-b256-31a9934c855d-ovn-node-metrics-cert\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245833 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245879 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-ovnkube-config\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245903 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-var-lib-openvswitch\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245923 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-slash\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245944 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-etc-openvswitch\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245996 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-cni-netd\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.246018 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-ovnkube-script-lib\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.245997 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.246052 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-cni-bin\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.246080 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-env-overrides\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.246122 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-openvswitch\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.246152 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-log-socket\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.246174 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-systemd\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.246198 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwkf6\" (UniqueName: \"kubernetes.io/projected/f89bfa10-15c0-40cd-b256-31a9934c855d-kube-api-access-rwkf6\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.246224 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-run-ovn-kubernetes\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.246892 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.248956 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.250011 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.251248 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.252288 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.255584 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.256158 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.257006 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.257725 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.258397 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.260056 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.260832 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.261893 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.261956 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.262840 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.265760 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.266447 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.267463 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.268102 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.270948 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.271770 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.296594 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.326405 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.327013 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-n44pt" event={"ID":"e908f308-e6e4-4a2e-ab9f-48e98698f2f7","Type":"ContainerStarted","Data":"2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4"} Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.327053 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-n44pt" event={"ID":"e908f308-e6e4-4a2e-ab9f-48e98698f2f7","Type":"ContainerStarted","Data":"6672a279fcfe21939852cd7cabd581482ab2d55e025f0efa981142bb26b0da0d"} Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.329109 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xq9dz" event={"ID":"aa83f85c-0050-4297-b65c-4837ecaf91b3","Type":"ContainerStarted","Data":"6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c"} Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.329134 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xq9dz" event={"ID":"aa83f85c-0050-4297-b65c-4837ecaf91b3","Type":"ContainerStarted","Data":"3f466dfb0829b69d75b608d3d673c6e0009256f7d01e22bdbccf2bbc64cfe642"} Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347201 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-etc-openvswitch\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347255 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-cni-netd\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347271 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-ovnkube-script-lib\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347301 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-cni-bin\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347315 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-env-overrides\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347334 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-openvswitch\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347354 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-log-socket\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347369 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-systemd\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347383 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwkf6\" (UniqueName: \"kubernetes.io/projected/f89bfa10-15c0-40cd-b256-31a9934c855d-kube-api-access-rwkf6\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347414 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-run-ovn-kubernetes\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347451 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-systemd-units\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347465 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-node-log\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347479 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-ovn\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347524 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-kubelet\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347547 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-run-netns\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347568 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f89bfa10-15c0-40cd-b256-31a9934c855d-ovn-node-metrics-cert\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347597 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347641 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-ovnkube-config\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347662 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-slash\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347682 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-var-lib-openvswitch\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347754 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-var-lib-openvswitch\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347804 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-etc-openvswitch\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.347839 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-cni-netd\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.348723 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-ovnkube-script-lib\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.348776 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-cni-bin\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.349181 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-env-overrides\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.349243 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-openvswitch\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.349274 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-log-socket\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.349304 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-systemd\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.349637 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-run-ovn-kubernetes\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.349686 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-systemd-units\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.349718 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-node-log\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.349744 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-ovn\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.349770 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-kubelet\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.349800 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-run-netns\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.355129 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-ovnkube-config\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.355196 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.355231 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-slash\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.363175 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f89bfa10-15c0-40cd-b256-31a9934c855d-ovn-node-metrics-cert\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.365092 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.378432 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwkf6\" (UniqueName: \"kubernetes.io/projected/f89bfa10-15c0-40cd-b256-31a9934c855d-kube-api-access-rwkf6\") pod \"ovnkube-node-lt499\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.396173 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.429057 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.447671 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.464427 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.476212 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: W1211 08:37:12.487950 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf89bfa10_15c0_40cd_b256_31a9934c855d.slice/crio-69a2b9f77a3183240a7feeadc3652a5779a90389c912b9dbd8b1b944421ae30c WatchSource:0}: Error finding container 69a2b9f77a3183240a7feeadc3652a5779a90389c912b9dbd8b1b944421ae30c: Status 404 returned error can't find the container with id 69a2b9f77a3183240a7feeadc3652a5779a90389c912b9dbd8b1b944421ae30c Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.507101 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.521860 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.540222 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.557482 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.574623 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.577273 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.593940 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.597523 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/847b50b9-0093-404f-b0dd-63c814d8986b-mcd-auth-proxy-config\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.609142 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.625067 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.639427 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.655173 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.673679 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.676918 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.677432 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-multus-daemon-config\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.697765 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.716376 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.734210 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.737220 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.738086 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-cni-binary-copy\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.738176 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0693c14e-aca1-4ada-8993-b17e90f51798-cni-binary-copy\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.749330 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.754682 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.778447 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.799721 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.818829 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.831395 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: E1211 08:37:12.837062 4629 secret.go:188] Couldn't get secret openshift-machine-config-operator/proxy-tls: failed to sync secret cache: timed out waiting for the condition Dec 11 08:37:12 crc kubenswrapper[4629]: E1211 08:37:12.837187 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/847b50b9-0093-404f-b0dd-63c814d8986b-proxy-tls podName:847b50b9-0093-404f-b0dd-63c814d8986b nodeName:}" failed. No retries permitted until 2025-12-11 08:37:13.337159525 +0000 UTC m=+21.402575133 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/847b50b9-0093-404f-b0dd-63c814d8986b-proxy-tls") pod "machine-config-daemon-wbld9" (UID: "847b50b9-0093-404f-b0dd-63c814d8986b") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.847411 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: E1211 08:37:12.861864 4629 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.866915 4629 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.868638 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.868683 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.868694 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.868761 4629 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.869272 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: E1211 08:37:12.872295 4629 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.877611 4629 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.878031 4629 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.879406 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.879479 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.879501 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.879524 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.879541 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:12Z","lastTransitionTime":"2025-12-11T08:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.884555 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.884751 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.895638 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-549hm\" (UniqueName: \"kubernetes.io/projected/847b50b9-0093-404f-b0dd-63c814d8986b-kube-api-access-549hm\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:12 crc kubenswrapper[4629]: E1211 08:37:12.902564 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.904924 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.906171 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.906210 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.906220 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.906236 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.906245 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:12Z","lastTransitionTime":"2025-12-11T08:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.917356 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: E1211 08:37:12.921014 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.926827 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.927089 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.927199 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.927315 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.927421 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:12Z","lastTransitionTime":"2025-12-11T08:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.936463 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: E1211 08:37:12.940162 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.943417 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.943456 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.943468 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.943492 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.943505 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:12Z","lastTransitionTime":"2025-12-11T08:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.949061 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 11 08:37:12 crc kubenswrapper[4629]: E1211 08:37:12.963128 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.966293 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.972775 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.973384 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.973494 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.973570 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.973636 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.973693 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:12Z","lastTransitionTime":"2025-12-11T08:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.981696 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.998981 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 11 08:37:12 crc kubenswrapper[4629]: I1211 08:37:12.999673 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.015387 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.015514 4629 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.017371 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.017401 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.017411 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.017427 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.017438 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:13Z","lastTransitionTime":"2025-12-11T08:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.026607 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.064160 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.119953 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.120325 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.120493 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.120589 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.120712 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:13Z","lastTransitionTime":"2025-12-11T08:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.136347 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.141972 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.170729 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.177296 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.197056 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.202537 4629 projected.go:194] Error preparing data for projected volume kube-api-access-xfwqs for pod openshift-multus/multus-lvjmx: failed to sync configmap cache: timed out waiting for the condition Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.202649 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-kube-api-access-xfwqs podName:f5568b3d-2e9b-4038-bb16-4ad3d87435ee nodeName:}" failed. No retries permitted until 2025-12-11 08:37:13.702626091 +0000 UTC m=+21.768041699 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-xfwqs" (UniqueName: "kubernetes.io/projected/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-kube-api-access-xfwqs") pod "multus-lvjmx" (UID: "f5568b3d-2e9b-4038-bb16-4ad3d87435ee") : failed to sync configmap cache: timed out waiting for the condition Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.202549 4629 projected.go:194] Error preparing data for projected volume kube-api-access-tqb45 for pod openshift-multus/multus-additional-cni-plugins-gn7n4: failed to sync configmap cache: timed out waiting for the condition Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.202691 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0693c14e-aca1-4ada-8993-b17e90f51798-kube-api-access-tqb45 podName:0693c14e-aca1-4ada-8993-b17e90f51798 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:13.702685243 +0000 UTC m=+21.768100851 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-tqb45" (UniqueName: "kubernetes.io/projected/0693c14e-aca1-4ada-8993-b17e90f51798-kube-api-access-tqb45") pod "multus-additional-cni-plugins-gn7n4" (UID: "0693c14e-aca1-4ada-8993-b17e90f51798") : failed to sync configmap cache: timed out waiting for the condition Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.213275 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.223298 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.223349 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.223362 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.223383 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.223395 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:13Z","lastTransitionTime":"2025-12-11T08:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.241740 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.262952 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.284316 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.305827 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.326250 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.326355 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.326369 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.326390 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.326404 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:13Z","lastTransitionTime":"2025-12-11T08:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.327540 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.332445 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272" exitCode=0 Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.332508 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272"} Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.332548 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"69a2b9f77a3183240a7feeadc3652a5779a90389c912b9dbd8b1b944421ae30c"} Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.333766 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b"} Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.355134 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.356501 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/847b50b9-0093-404f-b0dd-63c814d8986b-proxy-tls\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.360550 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/847b50b9-0093-404f-b0dd-63c814d8986b-proxy-tls\") pod \"machine-config-daemon-wbld9\" (UID: \"847b50b9-0093-404f-b0dd-63c814d8986b\") " pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.373340 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.392218 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.407224 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.421318 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.429180 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.429209 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.429219 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.429235 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.429244 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:13Z","lastTransitionTime":"2025-12-11T08:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.441563 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.457395 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.462480 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: W1211 08:37:13.469215 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod847b50b9_0093_404f_b0dd_63c814d8986b.slice/crio-61caa975e2f720d3b06037b2bab42fe2b14273d4a4f46210e530abf7ba9510b8 WatchSource:0}: Error finding container 61caa975e2f720d3b06037b2bab42fe2b14273d4a4f46210e530abf7ba9510b8: Status 404 returned error can't find the container with id 61caa975e2f720d3b06037b2bab42fe2b14273d4a4f46210e530abf7ba9510b8 Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.486176 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.532938 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.532979 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.532993 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.533018 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.533032 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:13Z","lastTransitionTime":"2025-12-11T08:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.533715 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.563578 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.604309 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.638954 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.638997 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.639008 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.639026 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.639039 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:13Z","lastTransitionTime":"2025-12-11T08:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.643084 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.683098 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.723829 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.741903 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.741940 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.741949 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.741964 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.741974 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:13Z","lastTransitionTime":"2025-12-11T08:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.761336 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfwqs\" (UniqueName: \"kubernetes.io/projected/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-kube-api-access-xfwqs\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.761399 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqb45\" (UniqueName: \"kubernetes.io/projected/0693c14e-aca1-4ada-8993-b17e90f51798-kube-api-access-tqb45\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.764995 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfwqs\" (UniqueName: \"kubernetes.io/projected/f5568b3d-2e9b-4038-bb16-4ad3d87435ee-kube-api-access-xfwqs\") pod \"multus-lvjmx\" (UID: \"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\") " pod="openshift-multus/multus-lvjmx" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.765290 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqb45\" (UniqueName: \"kubernetes.io/projected/0693c14e-aca1-4ada-8993-b17e90f51798-kube-api-access-tqb45\") pod \"multus-additional-cni-plugins-gn7n4\" (UID: \"0693c14e-aca1-4ada-8993-b17e90f51798\") " pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.766817 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:13Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.798067 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" Dec 11 08:37:13 crc kubenswrapper[4629]: W1211 08:37:13.824279 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0693c14e_aca1_4ada_8993_b17e90f51798.slice/crio-5486726ef4d4e1d1e287e60e7ab108c8f06ee7eea987696203785afefdd630a6 WatchSource:0}: Error finding container 5486726ef4d4e1d1e287e60e7ab108c8f06ee7eea987696203785afefdd630a6: Status 404 returned error can't find the container with id 5486726ef4d4e1d1e287e60e7ab108c8f06ee7eea987696203785afefdd630a6 Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.846792 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.846835 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.846867 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.846889 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.846901 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:13Z","lastTransitionTime":"2025-12-11T08:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.862789 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.862916 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.862948 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863014 4629 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863040 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:37:17.862999215 +0000 UTC m=+25.928414823 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863087 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:17.863076097 +0000 UTC m=+25.928491785 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.863142 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863175 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.863209 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863215 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863262 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863336 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863356 4629 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863267 4629 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863440 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:17.863418817 +0000 UTC m=+25.928834455 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863470 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:17.863457568 +0000 UTC m=+25.928873236 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863293 4629 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:13 crc kubenswrapper[4629]: E1211 08:37:13.863532 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:17.86351627 +0000 UTC m=+25.928931868 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.949644 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.949684 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.949698 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.949715 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:13 crc kubenswrapper[4629]: I1211 08:37:13.949727 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:13Z","lastTransitionTime":"2025-12-11T08:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.052068 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.052107 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.052119 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.052137 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.052149 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:14Z","lastTransitionTime":"2025-12-11T08:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.056633 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lvjmx" Dec 11 08:37:14 crc kubenswrapper[4629]: W1211 08:37:14.070545 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5568b3d_2e9b_4038_bb16_4ad3d87435ee.slice/crio-d8306acc093352555e43f4165eb91aa93accb88998b840a1a4823a5ec813000f WatchSource:0}: Error finding container d8306acc093352555e43f4165eb91aa93accb88998b840a1a4823a5ec813000f: Status 404 returned error can't find the container with id d8306acc093352555e43f4165eb91aa93accb88998b840a1a4823a5ec813000f Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.154587 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.154925 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.154944 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.154967 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.154979 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:14Z","lastTransitionTime":"2025-12-11T08:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.198361 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.198410 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:14 crc kubenswrapper[4629]: E1211 08:37:14.198489 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.198533 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:14 crc kubenswrapper[4629]: E1211 08:37:14.198550 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:14 crc kubenswrapper[4629]: E1211 08:37:14.198688 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.257232 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.257289 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.257302 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.257321 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.257333 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:14Z","lastTransitionTime":"2025-12-11T08:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.338857 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.338913 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.338926 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"61caa975e2f720d3b06037b2bab42fe2b14273d4a4f46210e530abf7ba9510b8"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.354178 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.354247 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.354257 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.354265 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.354274 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.354297 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.356438 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lvjmx" event={"ID":"f5568b3d-2e9b-4038-bb16-4ad3d87435ee","Type":"ContainerStarted","Data":"d8306acc093352555e43f4165eb91aa93accb88998b840a1a4823a5ec813000f"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.359599 4629 generic.go:334] "Generic (PLEG): container finished" podID="0693c14e-aca1-4ada-8993-b17e90f51798" containerID="2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7" exitCode=0 Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.359748 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" event={"ID":"0693c14e-aca1-4ada-8993-b17e90f51798","Type":"ContainerDied","Data":"2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.359793 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" event={"ID":"0693c14e-aca1-4ada-8993-b17e90f51798","Type":"ContainerStarted","Data":"5486726ef4d4e1d1e287e60e7ab108c8f06ee7eea987696203785afefdd630a6"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.360139 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.360186 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.360200 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.360219 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.360231 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:14Z","lastTransitionTime":"2025-12-11T08:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.362547 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.379523 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.393862 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.405758 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.423967 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.440424 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.458533 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.468817 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.468889 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.468900 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.468919 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.468951 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:14Z","lastTransitionTime":"2025-12-11T08:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.481800 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.500042 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.515448 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.536391 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.550923 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.563301 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.571230 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.571269 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.571280 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.571296 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.571305 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:14Z","lastTransitionTime":"2025-12-11T08:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.611519 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.644320 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.673660 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.673704 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.673720 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.673737 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.673748 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:14Z","lastTransitionTime":"2025-12-11T08:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.679317 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.697609 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.711263 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.725314 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.741179 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.759263 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.776376 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.776415 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.776425 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.776440 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.776451 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:14Z","lastTransitionTime":"2025-12-11T08:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.785463 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.805986 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.824876 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.845747 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.858447 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.871554 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.879679 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.879876 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.879957 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.880039 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.880100 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:14Z","lastTransitionTime":"2025-12-11T08:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.883682 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.983026 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.983084 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.983096 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.983114 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:14 crc kubenswrapper[4629]: I1211 08:37:14.983126 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:14Z","lastTransitionTime":"2025-12-11T08:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.085560 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.085790 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.085925 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.086025 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.086109 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:15Z","lastTransitionTime":"2025-12-11T08:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.188496 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.188556 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.188567 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.188585 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.188598 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:15Z","lastTransitionTime":"2025-12-11T08:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.291655 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.291693 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.291703 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.291719 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.291729 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:15Z","lastTransitionTime":"2025-12-11T08:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.365295 4629 generic.go:334] "Generic (PLEG): container finished" podID="0693c14e-aca1-4ada-8993-b17e90f51798" containerID="b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995" exitCode=0 Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.365369 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" event={"ID":"0693c14e-aca1-4ada-8993-b17e90f51798","Type":"ContainerDied","Data":"b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995"} Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.366697 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lvjmx" event={"ID":"f5568b3d-2e9b-4038-bb16-4ad3d87435ee","Type":"ContainerStarted","Data":"26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba"} Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.381415 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.394775 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.394815 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.394826 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.394857 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.394874 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:15Z","lastTransitionTime":"2025-12-11T08:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.401550 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.415558 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.430919 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.451662 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.464260 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.482616 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.500253 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.500288 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.500296 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.500312 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.500321 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:15Z","lastTransitionTime":"2025-12-11T08:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.502619 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.518556 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.535373 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.547613 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.557549 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.569839 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.581508 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.602903 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.602930 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.602938 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.602954 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.602963 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:15Z","lastTransitionTime":"2025-12-11T08:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.604623 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.617898 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.631163 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.644581 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.656576 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.681209 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.705270 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.705315 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.705329 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.705351 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.705371 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:15Z","lastTransitionTime":"2025-12-11T08:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.719633 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.764204 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.807409 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.808665 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.808717 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.808728 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.808745 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.808756 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:15Z","lastTransitionTime":"2025-12-11T08:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.841811 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.880901 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.911061 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.911104 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.911118 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.911137 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.911149 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:15Z","lastTransitionTime":"2025-12-11T08:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.917818 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:15 crc kubenswrapper[4629]: I1211 08:37:15.959718 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:15Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.003075 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.014075 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.014122 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.014135 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.014156 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.014169 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:16Z","lastTransitionTime":"2025-12-11T08:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.113103 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.117667 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.117714 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.117724 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.117740 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.117752 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:16Z","lastTransitionTime":"2025-12-11T08:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.123403 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.124636 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.129342 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.166746 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.198231 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.198259 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.198356 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:16 crc kubenswrapper[4629]: E1211 08:37:16.198401 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:16 crc kubenswrapper[4629]: E1211 08:37:16.198521 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:16 crc kubenswrapper[4629]: E1211 08:37:16.198621 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.209171 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.220722 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.220769 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.220780 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.220827 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.220855 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:16Z","lastTransitionTime":"2025-12-11T08:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.230983 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.250179 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.268139 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.298942 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.323511 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.323556 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.323568 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.323589 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.323601 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:16Z","lastTransitionTime":"2025-12-11T08:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.339501 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.371508 4629 generic.go:334] "Generic (PLEG): container finished" podID="0693c14e-aca1-4ada-8993-b17e90f51798" containerID="2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581" exitCode=0 Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.371573 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" event={"ID":"0693c14e-aca1-4ada-8993-b17e90f51798","Type":"ContainerDied","Data":"2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581"} Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.375004 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d"} Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.396058 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: E1211 08:37:16.402057 4629 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.426488 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.426532 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.426541 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.426560 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.426575 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:16Z","lastTransitionTime":"2025-12-11T08:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.441073 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.483599 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.521446 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.533259 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.533288 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.533296 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.533309 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.533319 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:16Z","lastTransitionTime":"2025-12-11T08:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.562468 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.609925 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.636975 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.637232 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.637246 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.637264 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.637278 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:16Z","lastTransitionTime":"2025-12-11T08:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.644436 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.681009 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.725198 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.740240 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.740299 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.740311 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.740332 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.740343 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:16Z","lastTransitionTime":"2025-12-11T08:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.757978 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.803061 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.843304 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.843578 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.843646 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.843709 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.843765 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:16Z","lastTransitionTime":"2025-12-11T08:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.849895 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.885380 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.924336 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.946229 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.946496 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.946606 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.946691 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.946761 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:16Z","lastTransitionTime":"2025-12-11T08:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:16 crc kubenswrapper[4629]: I1211 08:37:16.962248 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:16Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.002256 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.038200 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.048992 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.049265 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.049378 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.049452 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.049509 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:17Z","lastTransitionTime":"2025-12-11T08:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.088579 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.122533 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.161784 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.161819 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.161829 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.161858 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.161869 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:17Z","lastTransitionTime":"2025-12-11T08:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.174643 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.201659 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.264783 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.264886 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.264898 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.264912 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.264925 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:17Z","lastTransitionTime":"2025-12-11T08:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.368069 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.368114 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.368125 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.368145 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.368156 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:17Z","lastTransitionTime":"2025-12-11T08:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.381450 4629 generic.go:334] "Generic (PLEG): container finished" podID="0693c14e-aca1-4ada-8993-b17e90f51798" containerID="4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113" exitCode=0 Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.381561 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" event={"ID":"0693c14e-aca1-4ada-8993-b17e90f51798","Type":"ContainerDied","Data":"4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113"} Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.397631 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.424787 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.439200 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.454183 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.468730 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.471363 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.471386 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.471394 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.471408 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.471420 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:17Z","lastTransitionTime":"2025-12-11T08:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.482134 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.496751 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.520358 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.681701 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.685426 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.685598 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.685619 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.685647 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.685663 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:17Z","lastTransitionTime":"2025-12-11T08:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.696887 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.715599 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.728073 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.742463 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.762648 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.788548 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.788589 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.788599 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.788614 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.788625 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:17Z","lastTransitionTime":"2025-12-11T08:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.802897 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.883808 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.883888 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.883917 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.883946 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.883967 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.884073 4629 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.884154 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:25.884114346 +0000 UTC m=+33.949529954 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.884337 4629 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.884427 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:37:25.884397945 +0000 UTC m=+33.949813553 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.884472 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:25.884464327 +0000 UTC m=+33.949879935 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.884522 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.884564 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.884586 4629 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.884647 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:25.884628842 +0000 UTC m=+33.950044450 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.884797 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.884926 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.885031 4629 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:17 crc kubenswrapper[4629]: E1211 08:37:17.885131 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:25.885119758 +0000 UTC m=+33.950535416 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.892837 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.892904 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.892917 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.892935 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.892948 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:17Z","lastTransitionTime":"2025-12-11T08:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.996572 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.996718 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.996730 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.996744 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:17 crc kubenswrapper[4629]: I1211 08:37:17.996755 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:17Z","lastTransitionTime":"2025-12-11T08:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.099752 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.099816 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.099829 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.099870 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.099886 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:18Z","lastTransitionTime":"2025-12-11T08:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.198914 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.198988 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:18 crc kubenswrapper[4629]: E1211 08:37:18.199063 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:18 crc kubenswrapper[4629]: E1211 08:37:18.199175 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.199444 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:18 crc kubenswrapper[4629]: E1211 08:37:18.199734 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.202286 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.202322 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.202332 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.202348 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.202358 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:18Z","lastTransitionTime":"2025-12-11T08:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.306336 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.306560 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.306626 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.306695 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.306759 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:18Z","lastTransitionTime":"2025-12-11T08:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.388308 4629 generic.go:334] "Generic (PLEG): container finished" podID="0693c14e-aca1-4ada-8993-b17e90f51798" containerID="7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073" exitCode=0 Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.388388 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" event={"ID":"0693c14e-aca1-4ada-8993-b17e90f51798","Type":"ContainerDied","Data":"7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073"} Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.409497 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.409549 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.409557 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.409572 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.409582 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:18Z","lastTransitionTime":"2025-12-11T08:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.464876 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.483058 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.499474 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.513459 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.513877 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.513915 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.513931 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.513950 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.513963 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:18Z","lastTransitionTime":"2025-12-11T08:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.537049 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.553711 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.567580 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.584420 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.620709 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.620741 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.620751 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.620768 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.620780 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:18Z","lastTransitionTime":"2025-12-11T08:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.634217 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.668347 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.688680 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.703796 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.720411 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.726715 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.726766 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.726779 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.726797 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.726814 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:18Z","lastTransitionTime":"2025-12-11T08:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.742808 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.755652 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.829687 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.829721 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.829731 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.829746 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.829756 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:18Z","lastTransitionTime":"2025-12-11T08:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.932911 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.933205 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.933295 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.933390 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:18 crc kubenswrapper[4629]: I1211 08:37:18.933482 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:18Z","lastTransitionTime":"2025-12-11T08:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.037026 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.037098 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.037111 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.037131 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.037164 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:19Z","lastTransitionTime":"2025-12-11T08:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.140319 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.140359 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.140368 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.140383 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.140395 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:19Z","lastTransitionTime":"2025-12-11T08:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.242690 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.242730 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.242739 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.242754 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.242763 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:19Z","lastTransitionTime":"2025-12-11T08:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.345581 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.345624 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.345633 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.345647 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.345657 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:19Z","lastTransitionTime":"2025-12-11T08:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.394557 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb"} Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.394893 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.397222 4629 generic.go:334] "Generic (PLEG): container finished" podID="0693c14e-aca1-4ada-8993-b17e90f51798" containerID="038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90" exitCode=0 Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.397254 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" event={"ID":"0693c14e-aca1-4ada-8993-b17e90f51798","Type":"ContainerDied","Data":"038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90"} Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.415802 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.422986 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.430422 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.444138 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.448550 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.448606 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.448622 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.448646 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.448664 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:19Z","lastTransitionTime":"2025-12-11T08:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.459481 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.469821 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.482788 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.492951 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.503454 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.523335 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.537232 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.550942 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.552836 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.552972 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.552988 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.553004 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.553016 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:19Z","lastTransitionTime":"2025-12-11T08:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.561548 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.574253 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.587676 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.603836 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.615638 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.630409 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.641949 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.655148 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.655189 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.655201 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.655216 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.655226 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:19Z","lastTransitionTime":"2025-12-11T08:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.662087 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.676481 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.690705 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.705592 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.717038 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.730535 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.744572 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.755520 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.756872 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.756901 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.756912 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.756928 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.756939 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:19Z","lastTransitionTime":"2025-12-11T08:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.774661 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.799487 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.812685 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.826582 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.859639 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.859675 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.859683 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.859697 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.859707 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:19Z","lastTransitionTime":"2025-12-11T08:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.962535 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.962879 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.962989 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.963085 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:19 crc kubenswrapper[4629]: I1211 08:37:19.963167 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:19Z","lastTransitionTime":"2025-12-11T08:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.065880 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.065923 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.065934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.065952 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.065963 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:20Z","lastTransitionTime":"2025-12-11T08:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.168902 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.169317 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.169429 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.169514 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.169605 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:20Z","lastTransitionTime":"2025-12-11T08:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.198192 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.198300 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:20 crc kubenswrapper[4629]: E1211 08:37:20.198398 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.198471 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:20 crc kubenswrapper[4629]: E1211 08:37:20.198598 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:20 crc kubenswrapper[4629]: E1211 08:37:20.198677 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.272489 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.272544 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.272554 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.272573 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.272585 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:20Z","lastTransitionTime":"2025-12-11T08:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.379636 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.379709 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.379723 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.379742 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.379768 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:20Z","lastTransitionTime":"2025-12-11T08:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.407264 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" event={"ID":"0693c14e-aca1-4ada-8993-b17e90f51798","Type":"ContainerStarted","Data":"a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f"} Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.407367 4629 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.408008 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.426308 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.441058 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.446730 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.472046 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.483139 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.483176 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.483186 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.483201 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.483212 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:20Z","lastTransitionTime":"2025-12-11T08:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.486350 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.500708 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.514629 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.535301 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.548722 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.571520 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.586147 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.586422 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.586547 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.586636 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.586706 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:20Z","lastTransitionTime":"2025-12-11T08:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.588942 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.605784 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.619250 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.633506 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.652115 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.667996 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.684827 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.689771 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.689817 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.689831 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.689875 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.689889 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:20Z","lastTransitionTime":"2025-12-11T08:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.702103 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.720800 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.735314 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.747288 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.769981 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.787209 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.791774 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.791801 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.791809 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.791824 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.791833 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:20Z","lastTransitionTime":"2025-12-11T08:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.803125 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.818886 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.834307 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.849793 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.862650 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.875427 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.890455 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.894313 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.894367 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.894378 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.894397 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.894410 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:20Z","lastTransitionTime":"2025-12-11T08:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.909998 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:20Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.997240 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.997274 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.997285 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.997305 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:20 crc kubenswrapper[4629]: I1211 08:37:20.997317 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:20Z","lastTransitionTime":"2025-12-11T08:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.099980 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.100029 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.100041 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.100058 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.100068 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:21Z","lastTransitionTime":"2025-12-11T08:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.203487 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.203539 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.203549 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.203563 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.203574 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:21Z","lastTransitionTime":"2025-12-11T08:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.306833 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.306918 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.306930 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.306950 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.306964 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:21Z","lastTransitionTime":"2025-12-11T08:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.409926 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.409987 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.410003 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.410025 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.410038 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:21Z","lastTransitionTime":"2025-12-11T08:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.413708 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/0.log" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.418064 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb" exitCode=1 Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.418116 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb"} Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.419352 4629 scope.go:117] "RemoveContainer" containerID="130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.434946 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.453206 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.466388 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.479598 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.493363 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.506331 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.513745 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.514163 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.514174 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.514192 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.514204 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:21Z","lastTransitionTime":"2025-12-11T08:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.521545 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.543320 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:21Z\\\",\\\"message\\\":\\\"ry.go:117\\\\nI1211 08:37:20.777935 5820 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 08:37:20.778054 5820 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 08:37:20.778311 5820 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:37:20.778980 5820 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:37:20.779128 5820 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:37:20.779218 5820 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:37:20.779550 5820 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:37:20.780056 5820 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.557314 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.574984 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.602598 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.617149 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.617190 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.617201 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.617219 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.617236 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:21Z","lastTransitionTime":"2025-12-11T08:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.618329 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.634273 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.649512 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.661091 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.719624 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.719681 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.719696 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.719723 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.719742 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:21Z","lastTransitionTime":"2025-12-11T08:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.822083 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.822132 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.822144 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.822164 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.822179 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:21Z","lastTransitionTime":"2025-12-11T08:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.924678 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.924717 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.924726 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.924741 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:21 crc kubenswrapper[4629]: I1211 08:37:21.924752 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:21Z","lastTransitionTime":"2025-12-11T08:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.027570 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.027609 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.027617 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.027632 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.027642 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:22Z","lastTransitionTime":"2025-12-11T08:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.129733 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.129797 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.129809 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.129829 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.129841 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:22Z","lastTransitionTime":"2025-12-11T08:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.198369 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.198431 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:22 crc kubenswrapper[4629]: E1211 08:37:22.198537 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.198373 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:22 crc kubenswrapper[4629]: E1211 08:37:22.198713 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:22 crc kubenswrapper[4629]: E1211 08:37:22.198799 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.211326 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.223424 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.232197 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.232241 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.232254 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.232272 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.232288 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:22Z","lastTransitionTime":"2025-12-11T08:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.236290 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.251906 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.274215 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:21Z\\\",\\\"message\\\":\\\"ry.go:117\\\\nI1211 08:37:20.777935 5820 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 08:37:20.778054 5820 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 08:37:20.778311 5820 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:37:20.778980 5820 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:37:20.779128 5820 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:37:20.779218 5820 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:37:20.779550 5820 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:37:20.780056 5820 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.293352 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.309393 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.329712 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.335159 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.335200 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.335212 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.335227 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.335239 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:22Z","lastTransitionTime":"2025-12-11T08:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.345688 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.360520 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.375547 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.388777 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.401375 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.414027 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.425086 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/0.log" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.425177 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.428186 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363"} Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.428272 4629 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.437721 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.437774 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.437786 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.437804 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.437815 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:22Z","lastTransitionTime":"2025-12-11T08:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.443977 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.467368 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.482678 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.496119 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.509723 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.522230 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.533136 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.540827 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.540920 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.540937 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.541063 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.541084 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:22Z","lastTransitionTime":"2025-12-11T08:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.546086 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.560512 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.574506 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.594250 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:21Z\\\",\\\"message\\\":\\\"ry.go:117\\\\nI1211 08:37:20.777935 5820 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 08:37:20.778054 5820 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 08:37:20.778311 5820 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:37:20.778980 5820 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:37:20.779128 5820 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:37:20.779218 5820 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:37:20.779550 5820 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:37:20.780056 5820 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.612394 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.624677 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.638426 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.643413 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.643453 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.643465 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.643483 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.643493 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:22Z","lastTransitionTime":"2025-12-11T08:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.657439 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.745990 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.746046 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.746059 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.746080 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.746094 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:22Z","lastTransitionTime":"2025-12-11T08:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.849093 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.849150 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.849159 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.849177 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.849193 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:22Z","lastTransitionTime":"2025-12-11T08:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.953109 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.953165 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.953180 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.953214 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:22 crc kubenswrapper[4629]: I1211 08:37:22.953227 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:22Z","lastTransitionTime":"2025-12-11T08:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.021834 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.021916 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.021926 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.021946 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.021960 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: E1211 08:37:23.035373 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.039962 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.040027 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.040043 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.040068 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.040085 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: E1211 08:37:23.052957 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.058027 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.058080 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.058092 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.058110 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.058123 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: E1211 08:37:23.072862 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.077435 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.077489 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.077503 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.077522 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.077537 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: E1211 08:37:23.091963 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.096317 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.096362 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.096372 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.096387 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.096397 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: E1211 08:37:23.134108 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: E1211 08:37:23.134267 4629 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.136409 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.136534 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.136552 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.136572 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.136585 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.239021 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.239692 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.239709 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.239727 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.239739 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.342731 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.342791 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.342803 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.342823 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.342839 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.434215 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/1.log" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.434894 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/0.log" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.438624 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363" exitCode=1 Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.438689 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363"} Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.438761 4629 scope.go:117] "RemoveContainer" containerID="130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.439655 4629 scope.go:117] "RemoveContainer" containerID="2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363" Dec 11 08:37:23 crc kubenswrapper[4629]: E1211 08:37:23.439896 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.445863 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.445902 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.445915 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.445934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.445946 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.465994 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.487783 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.503540 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.519765 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.531388 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.543744 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.547940 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.547983 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.547993 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.548009 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.548021 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.555080 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.565023 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.582012 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://130f8e3750fbca6138adc037facad8840be86af4a07a309f1d19d01a067312cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:21Z\\\",\\\"message\\\":\\\"ry.go:117\\\\nI1211 08:37:20.777935 5820 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 08:37:20.778054 5820 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 08:37:20.778311 5820 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:37:20.778980 5820 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:37:20.779128 5820 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:37:20.779218 5820 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:37:20.779550 5820 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:37:20.780056 5820 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:22Z\\\",\\\"message\\\":\\\"d 0 (template) load balancers\\\\nI1211 08:37:22.244832 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:22.244880 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1211 08:37:22.244897 5978 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.594234 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.608992 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.619438 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.633195 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.643953 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.650695 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.650748 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.650760 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.650780 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.650792 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.659377 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:23Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.754265 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.754311 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.754320 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.754336 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.754345 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.856903 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.856950 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.856962 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.856976 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.856986 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.959892 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.959938 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.959947 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.959961 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:23 crc kubenswrapper[4629]: I1211 08:37:23.959973 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:23Z","lastTransitionTime":"2025-12-11T08:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.062746 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.062797 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.062809 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.062828 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.062858 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:24Z","lastTransitionTime":"2025-12-11T08:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.165826 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.165906 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.165923 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.165943 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.165956 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:24Z","lastTransitionTime":"2025-12-11T08:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.198943 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:24 crc kubenswrapper[4629]: E1211 08:37:24.199131 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.199190 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.199207 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:24 crc kubenswrapper[4629]: E1211 08:37:24.199356 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:24 crc kubenswrapper[4629]: E1211 08:37:24.199505 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.268194 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.268269 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.268279 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.268293 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.268304 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:24Z","lastTransitionTime":"2025-12-11T08:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.370750 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.370810 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.370834 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.370882 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.370897 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:24Z","lastTransitionTime":"2025-12-11T08:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.443519 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/1.log" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.447366 4629 scope.go:117] "RemoveContainer" containerID="2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363" Dec 11 08:37:24 crc kubenswrapper[4629]: E1211 08:37:24.447714 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.464175 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.473685 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.473718 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.473727 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.473742 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.473753 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:24Z","lastTransitionTime":"2025-12-11T08:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.480859 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.494637 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.507385 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.524769 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:22Z\\\",\\\"message\\\":\\\"d 0 (template) load balancers\\\\nI1211 08:37:22.244832 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:22.244880 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1211 08:37:22.244897 5978 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.537167 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.549209 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.565634 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.576271 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.576299 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.576308 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.576322 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.576331 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:24Z","lastTransitionTime":"2025-12-11T08:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.577148 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.590945 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.611465 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.626722 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.640873 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.655412 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.666309 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:24Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.678933 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.678959 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.678967 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.678980 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.678990 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:24Z","lastTransitionTime":"2025-12-11T08:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.781259 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.781300 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.781309 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.781325 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.781336 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:24Z","lastTransitionTime":"2025-12-11T08:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.884545 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.884595 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.884612 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.884633 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.884648 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:24Z","lastTransitionTime":"2025-12-11T08:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.987148 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.987202 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.987214 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.987236 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:24 crc kubenswrapper[4629]: I1211 08:37:24.987250 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:24Z","lastTransitionTime":"2025-12-11T08:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.090443 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.090494 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.090504 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.090522 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.090534 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:25Z","lastTransitionTime":"2025-12-11T08:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.193749 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.193820 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.193840 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.193890 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.193908 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:25Z","lastTransitionTime":"2025-12-11T08:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.297341 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.297383 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.297395 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.297414 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.297427 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:25Z","lastTransitionTime":"2025-12-11T08:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.399864 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.399921 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.399938 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.399962 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.399978 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:25Z","lastTransitionTime":"2025-12-11T08:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.494020 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4"] Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.494490 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:25 crc kubenswrapper[4629]: W1211 08:37:25.496714 4629 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert": failed to list *v1.Secret: secrets "ovn-control-plane-metrics-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.496831 4629 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-control-plane-metrics-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 11 08:37:25 crc kubenswrapper[4629]: W1211 08:37:25.496959 4629 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd": failed to list *v1.Secret: secrets "ovn-kubernetes-control-plane-dockercfg-gs7dd" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.496999 4629 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-gs7dd\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-control-plane-dockercfg-gs7dd\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.504366 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.504475 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.504494 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.504550 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.504569 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:25Z","lastTransitionTime":"2025-12-11T08:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.511360 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.524475 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.539466 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.552037 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.557669 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jvs5\" (UniqueName: \"kubernetes.io/projected/79696728-82a3-4065-b645-cfefc699b00f-kube-api-access-2jvs5\") pod \"ovnkube-control-plane-749d76644c-grzg4\" (UID: \"79696728-82a3-4065-b645-cfefc699b00f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.557755 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/79696728-82a3-4065-b645-cfefc699b00f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-grzg4\" (UID: \"79696728-82a3-4065-b645-cfefc699b00f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.557812 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/79696728-82a3-4065-b645-cfefc699b00f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-grzg4\" (UID: \"79696728-82a3-4065-b645-cfefc699b00f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.557838 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/79696728-82a3-4065-b645-cfefc699b00f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-grzg4\" (UID: \"79696728-82a3-4065-b645-cfefc699b00f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.570100 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.582191 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.592554 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.605036 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.607153 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.607198 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.607210 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.607231 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.607244 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:25Z","lastTransitionTime":"2025-12-11T08:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.623643 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.641439 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.652273 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.659362 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jvs5\" (UniqueName: \"kubernetes.io/projected/79696728-82a3-4065-b645-cfefc699b00f-kube-api-access-2jvs5\") pod \"ovnkube-control-plane-749d76644c-grzg4\" (UID: \"79696728-82a3-4065-b645-cfefc699b00f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.659453 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/79696728-82a3-4065-b645-cfefc699b00f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-grzg4\" (UID: \"79696728-82a3-4065-b645-cfefc699b00f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.659518 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/79696728-82a3-4065-b645-cfefc699b00f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-grzg4\" (UID: \"79696728-82a3-4065-b645-cfefc699b00f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.659544 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/79696728-82a3-4065-b645-cfefc699b00f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-grzg4\" (UID: \"79696728-82a3-4065-b645-cfefc699b00f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.665229 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/79696728-82a3-4065-b645-cfefc699b00f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-grzg4\" (UID: \"79696728-82a3-4065-b645-cfefc699b00f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.665691 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/79696728-82a3-4065-b645-cfefc699b00f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-grzg4\" (UID: \"79696728-82a3-4065-b645-cfefc699b00f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.668098 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.687703 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:22Z\\\",\\\"message\\\":\\\"d 0 (template) load balancers\\\\nI1211 08:37:22.244832 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:22.244880 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1211 08:37:22.244897 5978 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.690146 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jvs5\" (UniqueName: \"kubernetes.io/projected/79696728-82a3-4065-b645-cfefc699b00f-kube-api-access-2jvs5\") pod \"ovnkube-control-plane-749d76644c-grzg4\" (UID: \"79696728-82a3-4065-b645-cfefc699b00f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.705661 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.710549 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.710620 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.710635 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.710656 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.710691 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:25Z","lastTransitionTime":"2025-12-11T08:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.717910 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.730888 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.813309 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.813362 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.813404 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.813431 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.813449 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:25Z","lastTransitionTime":"2025-12-11T08:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.917998 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.918039 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.918050 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.918067 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.918079 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:25Z","lastTransitionTime":"2025-12-11T08:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.963677 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.963831 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.963943 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:37:41.963911552 +0000 UTC m=+50.029327160 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.963975 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.963997 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.964015 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:25 crc kubenswrapper[4629]: I1211 08:37:25.964031 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.964038 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.964052 4629 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.964069 4629 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.964098 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:41.964086878 +0000 UTC m=+50.029502486 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.964116 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:41.964106049 +0000 UTC m=+50.029521657 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.964118 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.964137 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.964146 4629 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.964163 4629 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.964176 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:41.964165021 +0000 UTC m=+50.029580629 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:25 crc kubenswrapper[4629]: E1211 08:37:25.964190 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:37:41.964183451 +0000 UTC m=+50.029599059 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.021239 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.021292 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.021302 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.021318 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.021328 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:26Z","lastTransitionTime":"2025-12-11T08:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.125076 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.125134 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.125150 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.125170 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.125186 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:26Z","lastTransitionTime":"2025-12-11T08:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.198145 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.198494 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:26 crc kubenswrapper[4629]: E1211 08:37:26.198774 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:26 crc kubenswrapper[4629]: E1211 08:37:26.198920 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.198794 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:26 crc kubenswrapper[4629]: E1211 08:37:26.199189 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.209106 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-29fct"] Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.209641 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:26 crc kubenswrapper[4629]: E1211 08:37:26.209715 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.224321 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.228465 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.228517 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.228530 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.228548 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.228561 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:26Z","lastTransitionTime":"2025-12-11T08:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.240366 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.256676 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.267911 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbz8h\" (UniqueName: \"kubernetes.io/projected/13e996ab-2163-44ad-9ecc-052a44f0da4c-kube-api-access-kbz8h\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.267953 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.271534 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.285162 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.294903 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.314900 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.330468 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.331052 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.331081 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.331099 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.331114 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.331124 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:26Z","lastTransitionTime":"2025-12-11T08:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.342969 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.356920 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.369066 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbz8h\" (UniqueName: \"kubernetes.io/projected/13e996ab-2163-44ad-9ecc-052a44f0da4c-kube-api-access-kbz8h\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.369130 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:26 crc kubenswrapper[4629]: E1211 08:37:26.369293 4629 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:26 crc kubenswrapper[4629]: E1211 08:37:26.369369 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs podName:13e996ab-2163-44ad-9ecc-052a44f0da4c nodeName:}" failed. No retries permitted until 2025-12-11 08:37:26.869349935 +0000 UTC m=+34.934765543 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs") pod "network-metrics-daemon-29fct" (UID: "13e996ab-2163-44ad-9ecc-052a44f0da4c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.371110 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.385332 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbz8h\" (UniqueName: \"kubernetes.io/projected/13e996ab-2163-44ad-9ecc-052a44f0da4c-kube-api-access-kbz8h\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.388475 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.402764 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.413126 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.424917 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.433256 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.433322 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.433339 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.433363 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.433419 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:26Z","lastTransitionTime":"2025-12-11T08:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.448405 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:22Z\\\",\\\"message\\\":\\\"d 0 (template) load balancers\\\\nI1211 08:37:22.244832 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:22.244880 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1211 08:37:22.244897 5978 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.460198 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:26Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.535904 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.536613 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.536710 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.536804 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.536899 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:26Z","lastTransitionTime":"2025-12-11T08:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.584147 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.595280 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/79696728-82a3-4065-b645-cfefc699b00f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-grzg4\" (UID: \"79696728-82a3-4065-b645-cfefc699b00f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.640122 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.640196 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.640214 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.640235 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.640244 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:26Z","lastTransitionTime":"2025-12-11T08:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.742460 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.742511 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.742520 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.742535 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.742544 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:26Z","lastTransitionTime":"2025-12-11T08:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.845321 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.845735 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.845759 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.845780 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.845791 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:26Z","lastTransitionTime":"2025-12-11T08:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.873810 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:26 crc kubenswrapper[4629]: E1211 08:37:26.874020 4629 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:26 crc kubenswrapper[4629]: E1211 08:37:26.874100 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs podName:13e996ab-2163-44ad-9ecc-052a44f0da4c nodeName:}" failed. No retries permitted until 2025-12-11 08:37:27.874079033 +0000 UTC m=+35.939494641 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs") pod "network-metrics-daemon-29fct" (UID: "13e996ab-2163-44ad-9ecc-052a44f0da4c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.948169 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.948233 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.948254 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.948274 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:26 crc kubenswrapper[4629]: I1211 08:37:26.948285 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:26Z","lastTransitionTime":"2025-12-11T08:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.013315 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.021285 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.050338 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.050389 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.050399 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.050415 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.050425 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:27Z","lastTransitionTime":"2025-12-11T08:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.153436 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.153469 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.153478 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.153492 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.153501 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:27Z","lastTransitionTime":"2025-12-11T08:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.255939 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.255994 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.256003 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.256018 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.256027 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:27Z","lastTransitionTime":"2025-12-11T08:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.277911 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.294052 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.308626 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.320391 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.334785 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.352834 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:22Z\\\",\\\"message\\\":\\\"d 0 (template) load balancers\\\\nI1211 08:37:22.244832 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:22.244880 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1211 08:37:22.244897 5978 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.358246 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.358279 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.358289 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.358304 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.358320 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:27Z","lastTransitionTime":"2025-12-11T08:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.364983 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.377319 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.394620 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.407430 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.432822 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.450110 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.456434 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" event={"ID":"79696728-82a3-4065-b645-cfefc699b00f","Type":"ContainerStarted","Data":"85f2ad8ec0930c09712e0df1fe75bb3e63c30bdbd2765ba9ed72ace59df91ac3"} Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.460204 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.460242 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.460256 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.460271 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.460282 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:27Z","lastTransitionTime":"2025-12-11T08:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.464599 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.476975 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.487629 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.499513 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.512743 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.523611 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.563139 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.563194 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.563208 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.563231 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.563246 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:27Z","lastTransitionTime":"2025-12-11T08:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.667243 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.667290 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.667323 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.667342 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.667354 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:27Z","lastTransitionTime":"2025-12-11T08:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.769826 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.769889 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.769899 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.769918 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.769945 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:27Z","lastTransitionTime":"2025-12-11T08:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.872932 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.872976 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.872986 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.873002 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.873014 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:27Z","lastTransitionTime":"2025-12-11T08:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.883802 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:27 crc kubenswrapper[4629]: E1211 08:37:27.883999 4629 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:27 crc kubenswrapper[4629]: E1211 08:37:27.884117 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs podName:13e996ab-2163-44ad-9ecc-052a44f0da4c nodeName:}" failed. No retries permitted until 2025-12-11 08:37:29.884091006 +0000 UTC m=+37.949506654 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs") pod "network-metrics-daemon-29fct" (UID: "13e996ab-2163-44ad-9ecc-052a44f0da4c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.975943 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.976016 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.976034 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.976063 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:27 crc kubenswrapper[4629]: I1211 08:37:27.976082 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:27Z","lastTransitionTime":"2025-12-11T08:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.079982 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.080061 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.080075 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.080095 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.080110 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:28Z","lastTransitionTime":"2025-12-11T08:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.184088 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.184170 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.184198 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.184282 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.184308 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:28Z","lastTransitionTime":"2025-12-11T08:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.198918 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.199080 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:28 crc kubenswrapper[4629]: E1211 08:37:28.199280 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:28 crc kubenswrapper[4629]: E1211 08:37:28.199115 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.199384 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.199574 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:28 crc kubenswrapper[4629]: E1211 08:37:28.199817 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:28 crc kubenswrapper[4629]: E1211 08:37:28.199941 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.288072 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.288118 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.288131 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.288149 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.288162 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:28Z","lastTransitionTime":"2025-12-11T08:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.391040 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.391077 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.391086 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.391100 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.391109 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:28Z","lastTransitionTime":"2025-12-11T08:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.494613 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.495000 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.495026 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.495044 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.495055 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:28Z","lastTransitionTime":"2025-12-11T08:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.598018 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.598086 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.598100 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.598120 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.598135 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:28Z","lastTransitionTime":"2025-12-11T08:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.702589 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.702646 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.702659 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.702681 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.702697 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:28Z","lastTransitionTime":"2025-12-11T08:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.806501 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.806584 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.806606 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.806634 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.806658 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:28Z","lastTransitionTime":"2025-12-11T08:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.910837 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.911273 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.911283 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.911301 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:28 crc kubenswrapper[4629]: I1211 08:37:28.911315 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:28Z","lastTransitionTime":"2025-12-11T08:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.013832 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.013900 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.013915 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.013934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.013949 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:29Z","lastTransitionTime":"2025-12-11T08:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.116864 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.116900 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.116911 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.116928 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.116940 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:29Z","lastTransitionTime":"2025-12-11T08:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.220074 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.220129 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.220139 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.220157 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.220171 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:29Z","lastTransitionTime":"2025-12-11T08:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.322973 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.323017 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.323028 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.323044 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.323056 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:29Z","lastTransitionTime":"2025-12-11T08:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.426099 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.426159 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.426169 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.426183 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.426193 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:29Z","lastTransitionTime":"2025-12-11T08:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.465899 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" event={"ID":"79696728-82a3-4065-b645-cfefc699b00f","Type":"ContainerStarted","Data":"8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1"} Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.465960 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" event={"ID":"79696728-82a3-4065-b645-cfefc699b00f","Type":"ContainerStarted","Data":"b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956"} Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.478046 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.495762 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.513160 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.533920 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.533977 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.533989 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.534008 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.534022 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:29Z","lastTransitionTime":"2025-12-11T08:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.539298 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.556828 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.570155 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.584700 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.596255 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.611749 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.628988 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.636631 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.636691 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.636714 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.636740 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.636757 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:29Z","lastTransitionTime":"2025-12-11T08:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.644980 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.659671 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.678620 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:22Z\\\",\\\"message\\\":\\\"d 0 (template) load balancers\\\\nI1211 08:37:22.244832 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:22.244880 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1211 08:37:22.244897 5978 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.697875 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.712104 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.726311 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.739965 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.740129 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.740153 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.740164 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.740181 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.740193 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:29Z","lastTransitionTime":"2025-12-11T08:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.842979 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.843102 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.843115 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.843130 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.843174 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:29Z","lastTransitionTime":"2025-12-11T08:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.909090 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:29 crc kubenswrapper[4629]: E1211 08:37:29.909304 4629 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:29 crc kubenswrapper[4629]: E1211 08:37:29.909438 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs podName:13e996ab-2163-44ad-9ecc-052a44f0da4c nodeName:}" failed. No retries permitted until 2025-12-11 08:37:33.909411821 +0000 UTC m=+41.974827519 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs") pod "network-metrics-daemon-29fct" (UID: "13e996ab-2163-44ad-9ecc-052a44f0da4c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.946572 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.946657 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.946702 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.946743 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:29 crc kubenswrapper[4629]: I1211 08:37:29.946771 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:29Z","lastTransitionTime":"2025-12-11T08:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.050001 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.050070 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.050091 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.050135 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.050176 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:30Z","lastTransitionTime":"2025-12-11T08:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.153401 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.153466 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.153475 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.153491 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.153501 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:30Z","lastTransitionTime":"2025-12-11T08:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.198649 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.198665 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:30 crc kubenswrapper[4629]: E1211 08:37:30.198878 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.198685 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.198665 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:30 crc kubenswrapper[4629]: E1211 08:37:30.198982 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:30 crc kubenswrapper[4629]: E1211 08:37:30.199055 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:30 crc kubenswrapper[4629]: E1211 08:37:30.199109 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.256598 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.256648 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.256659 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.256676 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.256687 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:30Z","lastTransitionTime":"2025-12-11T08:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.358974 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.359011 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.359020 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.359037 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.359047 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:30Z","lastTransitionTime":"2025-12-11T08:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.461308 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.461348 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.461358 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.461373 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.461384 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:30Z","lastTransitionTime":"2025-12-11T08:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.564428 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.564478 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.564490 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.564508 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.564520 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:30Z","lastTransitionTime":"2025-12-11T08:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.667561 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.667611 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.667628 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.667648 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.667660 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:30Z","lastTransitionTime":"2025-12-11T08:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.770154 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.770213 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.770224 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.770248 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.770259 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:30Z","lastTransitionTime":"2025-12-11T08:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.872743 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.872784 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.872796 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.872812 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.872821 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:30Z","lastTransitionTime":"2025-12-11T08:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.975287 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.975349 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.975362 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.975382 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:30 crc kubenswrapper[4629]: I1211 08:37:30.975396 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:30Z","lastTransitionTime":"2025-12-11T08:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.077953 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.077991 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.078001 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.078015 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.078024 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:31Z","lastTransitionTime":"2025-12-11T08:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.180895 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.180934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.180946 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.180964 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.180975 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:31Z","lastTransitionTime":"2025-12-11T08:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.284020 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.284065 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.284076 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.284102 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.284115 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:31Z","lastTransitionTime":"2025-12-11T08:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.387021 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.387051 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.387060 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.387074 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.387084 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:31Z","lastTransitionTime":"2025-12-11T08:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.489723 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.489774 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.489785 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.489800 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.489811 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:31Z","lastTransitionTime":"2025-12-11T08:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.592443 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.592702 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.592861 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.592953 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.593032 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:31Z","lastTransitionTime":"2025-12-11T08:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.696042 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.697120 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.697380 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.697576 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.697741 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:31Z","lastTransitionTime":"2025-12-11T08:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.800662 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.800698 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.800707 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.800722 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.800734 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:31Z","lastTransitionTime":"2025-12-11T08:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.903008 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.903318 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.903548 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.903644 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:31 crc kubenswrapper[4629]: I1211 08:37:31.903715 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:31Z","lastTransitionTime":"2025-12-11T08:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.007092 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.007153 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.007167 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.007189 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.007205 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:32Z","lastTransitionTime":"2025-12-11T08:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.110120 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.110169 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.110182 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.110201 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.110214 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:32Z","lastTransitionTime":"2025-12-11T08:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.199102 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.199142 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.199194 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.199290 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:32 crc kubenswrapper[4629]: E1211 08:37:32.200726 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:32 crc kubenswrapper[4629]: E1211 08:37:32.200871 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:32 crc kubenswrapper[4629]: E1211 08:37:32.200923 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:32 crc kubenswrapper[4629]: E1211 08:37:32.200999 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.213652 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.214136 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.214237 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.214347 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.214432 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:32Z","lastTransitionTime":"2025-12-11T08:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.214617 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.235481 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:22Z\\\",\\\"message\\\":\\\"d 0 (template) load balancers\\\\nI1211 08:37:22.244832 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:22.244880 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1211 08:37:22.244897 5978 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.247988 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.259676 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.271681 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.282365 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.293875 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.307144 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.317261 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.317606 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.317634 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.317927 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.317978 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:32Z","lastTransitionTime":"2025-12-11T08:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.319699 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.342886 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.357063 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.369890 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.382236 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.392762 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.404811 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.504623 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.507089 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.507141 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.507153 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.507175 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.507190 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:32Z","lastTransitionTime":"2025-12-11T08:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.517500 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.610093 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.610193 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.610205 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.610224 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.610251 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:32Z","lastTransitionTime":"2025-12-11T08:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.713072 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.713800 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.714055 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.714082 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.714093 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:32Z","lastTransitionTime":"2025-12-11T08:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.816943 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.817197 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.817285 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.817434 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.817510 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:32Z","lastTransitionTime":"2025-12-11T08:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.920434 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.920504 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.920517 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.920537 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:32 crc kubenswrapper[4629]: I1211 08:37:32.920564 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:32Z","lastTransitionTime":"2025-12-11T08:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.022453 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.022501 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.022515 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.022532 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.022545 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.125268 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.125331 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.125343 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.125377 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.125388 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.228393 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.228430 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.228441 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.228457 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.228468 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.331693 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.331744 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.331753 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.331767 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.331777 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.436123 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.436296 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.436321 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.436352 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.436372 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.467132 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.467202 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.467213 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.467232 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.467245 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: E1211 08:37:33.480686 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.485410 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.485473 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.485487 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.485508 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.485525 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: E1211 08:37:33.505301 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.511174 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.511235 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.511251 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.511271 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.511285 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: E1211 08:37:33.525508 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.529362 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.529402 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.529416 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.529435 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.529450 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: E1211 08:37:33.542362 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.547018 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.547059 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.547070 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.547087 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.547100 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: E1211 08:37:33.560154 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:33 crc kubenswrapper[4629]: E1211 08:37:33.560285 4629 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.562162 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.562194 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.562203 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.562218 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.562228 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.665431 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.665482 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.665515 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.665533 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.665545 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.768872 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.768923 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.768934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.768982 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.768995 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.872053 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.872293 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.872396 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.872470 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.872558 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.916293 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:33 crc kubenswrapper[4629]: E1211 08:37:33.916755 4629 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:33 crc kubenswrapper[4629]: E1211 08:37:33.916937 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs podName:13e996ab-2163-44ad-9ecc-052a44f0da4c nodeName:}" failed. No retries permitted until 2025-12-11 08:37:41.916915961 +0000 UTC m=+49.982331569 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs") pod "network-metrics-daemon-29fct" (UID: "13e996ab-2163-44ad-9ecc-052a44f0da4c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.976008 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.976065 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.976101 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.976121 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:33 crc kubenswrapper[4629]: I1211 08:37:33.976133 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:33Z","lastTransitionTime":"2025-12-11T08:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.078537 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.078868 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.078944 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.079043 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.079115 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:34Z","lastTransitionTime":"2025-12-11T08:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.181809 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.181876 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.181886 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.181902 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.181913 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:34Z","lastTransitionTime":"2025-12-11T08:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.198282 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.198282 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:34 crc kubenswrapper[4629]: E1211 08:37:34.198483 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.198298 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.198301 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:34 crc kubenswrapper[4629]: E1211 08:37:34.198642 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:34 crc kubenswrapper[4629]: E1211 08:37:34.198670 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:34 crc kubenswrapper[4629]: E1211 08:37:34.198735 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.285070 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.285113 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.285123 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.285138 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.285148 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:34Z","lastTransitionTime":"2025-12-11T08:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.387814 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.387868 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.387881 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.387901 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.387913 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:34Z","lastTransitionTime":"2025-12-11T08:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.491224 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.491274 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.491286 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.491305 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.491316 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:34Z","lastTransitionTime":"2025-12-11T08:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.594239 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.594320 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.594335 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.594353 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.594394 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:34Z","lastTransitionTime":"2025-12-11T08:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.697144 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.697187 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.697200 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.697216 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.697227 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:34Z","lastTransitionTime":"2025-12-11T08:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.800097 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.800137 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.800149 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.800168 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.800178 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:34Z","lastTransitionTime":"2025-12-11T08:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.902381 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.902420 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.902430 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.902444 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:34 crc kubenswrapper[4629]: I1211 08:37:34.902455 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:34Z","lastTransitionTime":"2025-12-11T08:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.005467 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.005512 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.005522 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.005539 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.005551 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:35Z","lastTransitionTime":"2025-12-11T08:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.108161 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.108204 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.108218 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.108244 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.108270 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:35Z","lastTransitionTime":"2025-12-11T08:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.211132 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.211236 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.211254 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.211278 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.211294 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:35Z","lastTransitionTime":"2025-12-11T08:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.314727 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.314924 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.314950 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.315005 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.315026 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:35Z","lastTransitionTime":"2025-12-11T08:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.417171 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.417245 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.417270 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.417299 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.417317 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:35Z","lastTransitionTime":"2025-12-11T08:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.519205 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.519250 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.519262 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.519282 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.519295 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:35Z","lastTransitionTime":"2025-12-11T08:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.622455 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.622502 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.622512 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.622530 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.622542 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:35Z","lastTransitionTime":"2025-12-11T08:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.725549 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.725602 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.725616 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.725638 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.725650 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:35Z","lastTransitionTime":"2025-12-11T08:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.828298 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.828357 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.828372 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.828393 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.828406 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:35Z","lastTransitionTime":"2025-12-11T08:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.930889 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.930948 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.930960 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.930979 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:35 crc kubenswrapper[4629]: I1211 08:37:35.930994 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:35Z","lastTransitionTime":"2025-12-11T08:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.033807 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.033872 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.033885 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.033906 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.033916 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:36Z","lastTransitionTime":"2025-12-11T08:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.136056 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.136118 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.136138 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.136156 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.136168 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:36Z","lastTransitionTime":"2025-12-11T08:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.197935 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.197990 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.197960 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.197935 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:36 crc kubenswrapper[4629]: E1211 08:37:36.198118 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:36 crc kubenswrapper[4629]: E1211 08:37:36.198186 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:36 crc kubenswrapper[4629]: E1211 08:37:36.198354 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:36 crc kubenswrapper[4629]: E1211 08:37:36.198474 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.240105 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.240193 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.240218 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.240251 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.240277 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:36Z","lastTransitionTime":"2025-12-11T08:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.342830 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.343134 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.343271 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.343432 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.343550 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:36Z","lastTransitionTime":"2025-12-11T08:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.445428 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.445470 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.445481 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.445498 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.445508 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:36Z","lastTransitionTime":"2025-12-11T08:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.547974 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.548203 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.548270 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.548333 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.548388 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:36Z","lastTransitionTime":"2025-12-11T08:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.650759 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.650814 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.650824 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.650873 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.650884 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:36Z","lastTransitionTime":"2025-12-11T08:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.753742 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.753833 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.753875 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.753899 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.753946 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:36Z","lastTransitionTime":"2025-12-11T08:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.856651 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.857321 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.857436 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.857529 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.857653 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:36Z","lastTransitionTime":"2025-12-11T08:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.960533 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.960596 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.960620 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.960649 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:36 crc kubenswrapper[4629]: I1211 08:37:36.960671 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:36Z","lastTransitionTime":"2025-12-11T08:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.063545 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.063968 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.064178 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.064326 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.064460 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:37Z","lastTransitionTime":"2025-12-11T08:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.167547 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.168209 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.168582 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.168704 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.168789 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:37Z","lastTransitionTime":"2025-12-11T08:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.271413 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.271755 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.271896 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.271999 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.272100 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:37Z","lastTransitionTime":"2025-12-11T08:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.374894 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.374931 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.374939 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.374955 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.374965 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:37Z","lastTransitionTime":"2025-12-11T08:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.477834 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.477895 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.477902 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.477918 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.477927 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:37Z","lastTransitionTime":"2025-12-11T08:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.581041 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.581404 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.581475 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.581584 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.581658 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:37Z","lastTransitionTime":"2025-12-11T08:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.684614 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.684655 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.684665 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.684682 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.684693 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:37Z","lastTransitionTime":"2025-12-11T08:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.786917 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.786974 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.786988 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.787006 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.787018 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:37Z","lastTransitionTime":"2025-12-11T08:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.890215 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.890870 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.891002 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.891102 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.891323 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:37Z","lastTransitionTime":"2025-12-11T08:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.993468 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.993740 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.993828 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.993947 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:37 crc kubenswrapper[4629]: I1211 08:37:37.994033 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:37Z","lastTransitionTime":"2025-12-11T08:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.097955 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.098008 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.098019 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.098051 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.098102 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:38Z","lastTransitionTime":"2025-12-11T08:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.198249 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.198312 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.198489 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.198526 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:38 crc kubenswrapper[4629]: E1211 08:37:38.198548 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:38 crc kubenswrapper[4629]: E1211 08:37:38.198661 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:38 crc kubenswrapper[4629]: E1211 08:37:38.198819 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:38 crc kubenswrapper[4629]: E1211 08:37:38.198907 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.200289 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.200413 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.200495 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.200582 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.200666 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:38Z","lastTransitionTime":"2025-12-11T08:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.303732 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.304077 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.304158 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.304273 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.304351 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:38Z","lastTransitionTime":"2025-12-11T08:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.407381 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.407711 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.407793 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.407905 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.407997 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:38Z","lastTransitionTime":"2025-12-11T08:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.510863 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.510898 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.510909 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.510927 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.510939 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:38Z","lastTransitionTime":"2025-12-11T08:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.613935 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.613974 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.613984 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.614003 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.614014 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:38Z","lastTransitionTime":"2025-12-11T08:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.716609 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.716659 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.716669 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.716688 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.716699 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:38Z","lastTransitionTime":"2025-12-11T08:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.819588 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.819626 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.819638 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.819656 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.819668 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:38Z","lastTransitionTime":"2025-12-11T08:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.922400 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.922456 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.922471 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.922489 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:38 crc kubenswrapper[4629]: I1211 08:37:38.922500 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:38Z","lastTransitionTime":"2025-12-11T08:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.025025 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.025345 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.025411 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.025481 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.025540 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:39Z","lastTransitionTime":"2025-12-11T08:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.086300 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.097535 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.105886 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.120213 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.128236 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.128279 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.128291 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.128309 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.128322 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:39Z","lastTransitionTime":"2025-12-11T08:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.137254 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.148340 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.167947 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.181375 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.192648 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.199268 4629 scope.go:117] "RemoveContainer" containerID="2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.207303 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.219061 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.230382 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.230431 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.230443 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.230460 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.230470 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:39Z","lastTransitionTime":"2025-12-11T08:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.234880 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.244194 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.331325 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.333178 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.333207 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.333216 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.333231 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.333241 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:39Z","lastTransitionTime":"2025-12-11T08:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.356628 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:22Z\\\",\\\"message\\\":\\\"d 0 (template) load balancers\\\\nI1211 08:37:22.244832 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:22.244880 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1211 08:37:22.244897 5978 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.368389 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.384500 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.398363 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.417404 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.435221 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.435267 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.435277 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.435294 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.435305 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:39Z","lastTransitionTime":"2025-12-11T08:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.534044 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/1.log" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.536756 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.536800 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.536810 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.536829 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.536855 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:39Z","lastTransitionTime":"2025-12-11T08:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.537647 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758"} Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.537955 4629 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.552907 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.568085 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.580705 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.605020 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.627068 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.640649 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.640691 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.640704 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.640722 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.640735 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:39Z","lastTransitionTime":"2025-12-11T08:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.651954 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.671258 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.693943 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.719975 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.737346 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.742934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.742978 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.742986 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.743001 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.743013 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:39Z","lastTransitionTime":"2025-12-11T08:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.751477 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.767807 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.784836 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:22Z\\\",\\\"message\\\":\\\"d 0 (template) load balancers\\\\nI1211 08:37:22.244832 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:22.244880 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1211 08:37:22.244897 5978 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.796068 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.808990 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.822731 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.835539 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.845657 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.845689 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.845698 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.845712 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.845721 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:39Z","lastTransitionTime":"2025-12-11T08:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.846372 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.948141 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.948260 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.948271 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.948286 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:39 crc kubenswrapper[4629]: I1211 08:37:39.948297 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:39Z","lastTransitionTime":"2025-12-11T08:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.050284 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.050318 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.050327 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.050343 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.050353 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:40Z","lastTransitionTime":"2025-12-11T08:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.153431 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.153490 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.153510 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.153536 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.153554 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:40Z","lastTransitionTime":"2025-12-11T08:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.198863 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.198913 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.198942 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.198863 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:40 crc kubenswrapper[4629]: E1211 08:37:40.199032 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:40 crc kubenswrapper[4629]: E1211 08:37:40.199099 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:40 crc kubenswrapper[4629]: E1211 08:37:40.199156 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:40 crc kubenswrapper[4629]: E1211 08:37:40.199207 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.257010 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.257075 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.257087 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.257107 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.257120 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:40Z","lastTransitionTime":"2025-12-11T08:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.359546 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.359613 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.359624 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.359643 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.359655 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:40Z","lastTransitionTime":"2025-12-11T08:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.462218 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.462258 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.462271 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.462289 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.462300 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:40Z","lastTransitionTime":"2025-12-11T08:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.542650 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/2.log" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.543249 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/1.log" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.546017 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758" exitCode=1 Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.546056 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758"} Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.546109 4629 scope.go:117] "RemoveContainer" containerID="2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.547069 4629 scope.go:117] "RemoveContainer" containerID="3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758" Dec 11 08:37:40 crc kubenswrapper[4629]: E1211 08:37:40.549032 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.563228 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.565914 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.565947 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.565955 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.565968 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.565977 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:40Z","lastTransitionTime":"2025-12-11T08:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.577662 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.588636 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.606930 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.621523 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.633475 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.647746 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.657775 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.668443 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.668512 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.668522 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.668540 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.668551 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:40Z","lastTransitionTime":"2025-12-11T08:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.670288 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.682949 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.693900 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.706804 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.725759 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c828c6f4a3638ea5955fa89592b47b561a48e066ba7dd97e5fe877bac5aa363\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:22Z\\\",\\\"message\\\":\\\"d 0 (template) load balancers\\\\nI1211 08:37:22.244832 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:22.244880 5978 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1211 08:37:22.244897 5978 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:40Z\\\",\\\"message\\\":\\\"93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:40.084266 6197 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:37:40.084173 6197 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 08:37:40.084328 6197 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nF1211 08:37:40.084345 6197 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.738307 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.750039 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.751990 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.761409 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.770788 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.770827 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.770839 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.770877 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.770891 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:40Z","lastTransitionTime":"2025-12-11T08:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.774091 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.784178 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.874045 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.874090 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.874100 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.874116 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.874339 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:40Z","lastTransitionTime":"2025-12-11T08:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.976993 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.977031 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.977040 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.977057 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:40 crc kubenswrapper[4629]: I1211 08:37:40.977068 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:40Z","lastTransitionTime":"2025-12-11T08:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.079360 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.079422 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.079441 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.079464 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.079480 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:41Z","lastTransitionTime":"2025-12-11T08:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.182699 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.182758 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.182767 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.182786 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.182796 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:41Z","lastTransitionTime":"2025-12-11T08:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.285220 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.285264 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.285275 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.285291 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.285299 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:41Z","lastTransitionTime":"2025-12-11T08:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.388632 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.388679 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.388690 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.388709 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.388727 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:41Z","lastTransitionTime":"2025-12-11T08:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.491099 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.491139 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.491147 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.491161 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.491172 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:41Z","lastTransitionTime":"2025-12-11T08:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.553063 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/2.log" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.556518 4629 scope.go:117] "RemoveContainer" containerID="3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758" Dec 11 08:37:41 crc kubenswrapper[4629]: E1211 08:37:41.556945 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.568792 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.584515 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.594229 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.594274 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.594323 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.594343 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.594356 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:41Z","lastTransitionTime":"2025-12-11T08:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.602822 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:40Z\\\",\\\"message\\\":\\\"93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:40.084266 6197 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:37:40.084173 6197 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 08:37:40.084328 6197 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nF1211 08:37:40.084345 6197 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.614562 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.627062 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.640744 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.653360 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.664814 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.681684 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.697070 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.697477 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.697512 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.697524 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.697544 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.697557 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:41Z","lastTransitionTime":"2025-12-11T08:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.717419 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.742286 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.756585 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.772803 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.786913 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.800632 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.800700 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.800714 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.800731 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.800742 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:41Z","lastTransitionTime":"2025-12-11T08:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.800925 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.814007 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.826718 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.903646 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.903703 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.903716 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.903737 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.903750 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:41Z","lastTransitionTime":"2025-12-11T08:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:41 crc kubenswrapper[4629]: I1211 08:37:41.952159 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:41 crc kubenswrapper[4629]: E1211 08:37:41.952404 4629 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:41 crc kubenswrapper[4629]: E1211 08:37:41.952667 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs podName:13e996ab-2163-44ad-9ecc-052a44f0da4c nodeName:}" failed. No retries permitted until 2025-12-11 08:37:57.952641734 +0000 UTC m=+66.018057402 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs") pod "network-metrics-daemon-29fct" (UID: "13e996ab-2163-44ad-9ecc-052a44f0da4c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.007809 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.007892 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.007903 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.007921 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.007937 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:42Z","lastTransitionTime":"2025-12-11T08:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.052509 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.052602 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.052640 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.052684 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.052712 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.052795 4629 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.052817 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.052871 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.052883 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:38:14.052821557 +0000 UTC m=+82.118237165 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.052938 4629 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.052963 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.053023 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.053037 4629 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.052899 4629 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.052983 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:38:14.052973032 +0000 UTC m=+82.118388640 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.053191 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:38:14.053181679 +0000 UTC m=+82.118597287 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.053207 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:38:14.053199829 +0000 UTC m=+82.118615437 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.053218 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:38:14.05321303 +0000 UTC m=+82.118628638 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.110779 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.110839 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.110877 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.110898 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.110912 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:42Z","lastTransitionTime":"2025-12-11T08:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.198424 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.198466 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.198467 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.198448 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.198595 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.198708 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.198787 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:42 crc kubenswrapper[4629]: E1211 08:37:42.198830 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.209703 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.218317 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.218368 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.218386 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.218410 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.218425 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:42Z","lastTransitionTime":"2025-12-11T08:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.231624 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.243059 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.256910 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.267446 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.291329 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.304693 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.319959 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.321541 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.321684 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.321762 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.321864 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.321952 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:42Z","lastTransitionTime":"2025-12-11T08:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.334427 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.349651 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.360605 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.372768 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.383621 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.395729 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.414264 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:40Z\\\",\\\"message\\\":\\\"93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:40.084266 6197 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:37:40.084173 6197 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 08:37:40.084328 6197 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nF1211 08:37:40.084345 6197 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.425325 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.425388 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.425413 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.425438 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.425451 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:42Z","lastTransitionTime":"2025-12-11T08:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.429506 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.444265 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.459309 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.527606 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.527636 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.527644 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.527659 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.527668 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:42Z","lastTransitionTime":"2025-12-11T08:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.631018 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.631074 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.631088 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.631107 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.631120 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:42Z","lastTransitionTime":"2025-12-11T08:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.734218 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.734278 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.734295 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.734315 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.734327 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:42Z","lastTransitionTime":"2025-12-11T08:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.836612 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.836666 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.836676 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.836695 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.836706 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:42Z","lastTransitionTime":"2025-12-11T08:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.939614 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.939735 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.939750 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.939791 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:42 crc kubenswrapper[4629]: I1211 08:37:42.939815 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:42Z","lastTransitionTime":"2025-12-11T08:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.042799 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.042906 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.042922 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.042946 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.042965 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.145424 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.145464 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.145476 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.145503 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.145515 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.248326 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.248375 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.248387 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.248405 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.248422 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.351744 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.351784 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.351795 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.351812 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.351823 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.454385 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.454667 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.454699 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.454718 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.454732 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.564969 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.565010 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.565022 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.565037 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.565049 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.667250 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.667326 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.667337 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.667359 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.667378 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.759209 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.759255 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.759267 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.759285 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.759296 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: E1211 08:37:43.771243 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.775204 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.775235 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.775244 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.775261 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.775271 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: E1211 08:37:43.785966 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.789794 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.789829 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.789868 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.789888 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.789900 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: E1211 08:37:43.802356 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.809132 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.809176 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.809186 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.809203 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.809217 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: E1211 08:37:43.820658 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.824264 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.824307 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.824319 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.824343 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.824356 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: E1211 08:37:43.836581 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:43 crc kubenswrapper[4629]: E1211 08:37:43.836744 4629 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.838223 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.838252 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.838262 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.838279 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.838288 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.940707 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.940743 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.940751 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.940765 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:43 crc kubenswrapper[4629]: I1211 08:37:43.940774 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:43Z","lastTransitionTime":"2025-12-11T08:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.043524 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.043571 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.043580 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.043598 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.043608 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:44Z","lastTransitionTime":"2025-12-11T08:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.146890 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.146954 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.146964 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.146982 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.146993 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:44Z","lastTransitionTime":"2025-12-11T08:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.198730 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.198722 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.198779 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.198748 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:44 crc kubenswrapper[4629]: E1211 08:37:44.198932 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:44 crc kubenswrapper[4629]: E1211 08:37:44.199093 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:44 crc kubenswrapper[4629]: E1211 08:37:44.199173 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:44 crc kubenswrapper[4629]: E1211 08:37:44.199236 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.249384 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.249430 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.249445 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.249463 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.249475 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:44Z","lastTransitionTime":"2025-12-11T08:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.352180 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.352219 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.352229 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.352250 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.352266 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:44Z","lastTransitionTime":"2025-12-11T08:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.455366 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.455417 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.455431 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.455451 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.455465 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:44Z","lastTransitionTime":"2025-12-11T08:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.558237 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.558284 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.558293 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.558307 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.558318 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:44Z","lastTransitionTime":"2025-12-11T08:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.660356 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.660401 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.660412 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.660429 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.660439 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:44Z","lastTransitionTime":"2025-12-11T08:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.762898 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.762973 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.762991 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.763018 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.763045 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:44Z","lastTransitionTime":"2025-12-11T08:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.865268 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.865339 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.865359 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.865380 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.865392 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:44Z","lastTransitionTime":"2025-12-11T08:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.967536 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.967578 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.967589 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.967608 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:44 crc kubenswrapper[4629]: I1211 08:37:44.967623 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:44Z","lastTransitionTime":"2025-12-11T08:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.073579 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.073615 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.073654 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.073883 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.073900 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:45Z","lastTransitionTime":"2025-12-11T08:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.176372 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.176417 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.176427 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.176444 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.176454 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:45Z","lastTransitionTime":"2025-12-11T08:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.278834 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.279142 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.279237 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.279334 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.279400 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:45Z","lastTransitionTime":"2025-12-11T08:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.383019 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.383048 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.383056 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.383071 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.383080 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:45Z","lastTransitionTime":"2025-12-11T08:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.485999 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.486565 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.486576 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.486594 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.486605 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:45Z","lastTransitionTime":"2025-12-11T08:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.588536 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.588586 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.588595 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.588612 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.588624 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:45Z","lastTransitionTime":"2025-12-11T08:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.692072 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.692129 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.692145 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.692167 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.692180 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:45Z","lastTransitionTime":"2025-12-11T08:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.795898 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.796025 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.796034 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.796051 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.796062 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:45Z","lastTransitionTime":"2025-12-11T08:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.898552 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.898599 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.898612 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.898635 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:45 crc kubenswrapper[4629]: I1211 08:37:45.898648 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:45Z","lastTransitionTime":"2025-12-11T08:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.001356 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.001409 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.001421 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.001467 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.001481 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:46Z","lastTransitionTime":"2025-12-11T08:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.104838 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.104914 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.104926 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.104946 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.104960 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:46Z","lastTransitionTime":"2025-12-11T08:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.198203 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.198259 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.198290 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:46 crc kubenswrapper[4629]: E1211 08:37:46.198387 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.198426 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:46 crc kubenswrapper[4629]: E1211 08:37:46.198550 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:46 crc kubenswrapper[4629]: E1211 08:37:46.198633 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:46 crc kubenswrapper[4629]: E1211 08:37:46.198711 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.207089 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.207129 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.207143 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.207158 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.207170 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:46Z","lastTransitionTime":"2025-12-11T08:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.309463 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.309520 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.309532 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.309553 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.309566 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:46Z","lastTransitionTime":"2025-12-11T08:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.413149 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.413212 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.413225 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.413245 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.413260 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:46Z","lastTransitionTime":"2025-12-11T08:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.515577 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.515629 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.515638 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.515655 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.515665 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:46Z","lastTransitionTime":"2025-12-11T08:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.617990 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.618060 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.618069 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.618085 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.618095 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:46Z","lastTransitionTime":"2025-12-11T08:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.721154 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.721225 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.721243 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.721565 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.721605 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:46Z","lastTransitionTime":"2025-12-11T08:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.825301 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.825362 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.825380 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.825401 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.825423 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:46Z","lastTransitionTime":"2025-12-11T08:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.928472 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.928515 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.928529 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.928545 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:46 crc kubenswrapper[4629]: I1211 08:37:46.928557 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:46Z","lastTransitionTime":"2025-12-11T08:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.031316 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.031381 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.031393 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.031412 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.031425 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:47Z","lastTransitionTime":"2025-12-11T08:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.134341 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.134396 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.134411 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.134427 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.134438 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:47Z","lastTransitionTime":"2025-12-11T08:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.237398 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.237443 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.237460 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.237482 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.237500 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:47Z","lastTransitionTime":"2025-12-11T08:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.340277 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.340322 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.340333 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.340350 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.340360 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:47Z","lastTransitionTime":"2025-12-11T08:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.442428 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.442474 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.442487 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.442503 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.442516 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:47Z","lastTransitionTime":"2025-12-11T08:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.550616 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.550670 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.550683 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.550700 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.550713 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:47Z","lastTransitionTime":"2025-12-11T08:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.653549 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.653614 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.653627 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.653652 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.653666 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:47Z","lastTransitionTime":"2025-12-11T08:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.756251 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.756305 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.756319 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.756338 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.756350 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:47Z","lastTransitionTime":"2025-12-11T08:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.859791 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.859865 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.859879 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.859896 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.859907 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:47Z","lastTransitionTime":"2025-12-11T08:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.962699 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.962749 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.962762 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.962780 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:47 crc kubenswrapper[4629]: I1211 08:37:47.962793 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:47Z","lastTransitionTime":"2025-12-11T08:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.065649 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.065693 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.065704 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.065719 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.065730 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:48Z","lastTransitionTime":"2025-12-11T08:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.167815 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.167872 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.167886 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.167906 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.167918 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:48Z","lastTransitionTime":"2025-12-11T08:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.198365 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.198403 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.198403 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.198427 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:48 crc kubenswrapper[4629]: E1211 08:37:48.198534 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:48 crc kubenswrapper[4629]: E1211 08:37:48.198808 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:48 crc kubenswrapper[4629]: E1211 08:37:48.199205 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:48 crc kubenswrapper[4629]: E1211 08:37:48.199277 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.270884 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.270938 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.270952 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.270973 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.270988 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:48Z","lastTransitionTime":"2025-12-11T08:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.374063 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.374107 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.374120 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.374138 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.374150 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:48Z","lastTransitionTime":"2025-12-11T08:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.475974 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.476014 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.476027 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.476040 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.476050 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:48Z","lastTransitionTime":"2025-12-11T08:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.578787 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.578821 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.578834 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.578875 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.578888 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:48Z","lastTransitionTime":"2025-12-11T08:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.681800 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.681840 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.681871 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.681886 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.681898 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:48Z","lastTransitionTime":"2025-12-11T08:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.784085 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.784135 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.784147 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.784165 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.784179 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:48Z","lastTransitionTime":"2025-12-11T08:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.886813 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.886880 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.886890 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.886904 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.886916 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:48Z","lastTransitionTime":"2025-12-11T08:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.989694 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.989738 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.989748 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.989764 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:48 crc kubenswrapper[4629]: I1211 08:37:48.989775 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:48Z","lastTransitionTime":"2025-12-11T08:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.092786 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.092869 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.092882 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.092906 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.092924 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:49Z","lastTransitionTime":"2025-12-11T08:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.195590 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.195635 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.195647 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.195665 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.195677 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:49Z","lastTransitionTime":"2025-12-11T08:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.298705 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.298747 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.298755 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.298774 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.298788 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:49Z","lastTransitionTime":"2025-12-11T08:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.402190 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.402251 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.402265 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.402285 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.402304 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:49Z","lastTransitionTime":"2025-12-11T08:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.504206 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.504245 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.504257 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.504274 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.504283 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:49Z","lastTransitionTime":"2025-12-11T08:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.607388 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.607430 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.607441 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.607460 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.607472 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:49Z","lastTransitionTime":"2025-12-11T08:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.711430 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.711474 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.711485 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.711500 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.711510 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:49Z","lastTransitionTime":"2025-12-11T08:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.814063 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.814110 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.814121 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.814139 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.814150 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:49Z","lastTransitionTime":"2025-12-11T08:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.916773 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.916877 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.916891 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.916910 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:49 crc kubenswrapper[4629]: I1211 08:37:49.916928 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:49Z","lastTransitionTime":"2025-12-11T08:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.019649 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.019692 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.019702 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.019717 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.019727 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:50Z","lastTransitionTime":"2025-12-11T08:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.121717 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.121791 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.121815 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.121921 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.121960 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:50Z","lastTransitionTime":"2025-12-11T08:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.198663 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.198708 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.198798 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.198663 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:50 crc kubenswrapper[4629]: E1211 08:37:50.198889 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:50 crc kubenswrapper[4629]: E1211 08:37:50.198956 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:50 crc kubenswrapper[4629]: E1211 08:37:50.199034 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:50 crc kubenswrapper[4629]: E1211 08:37:50.199092 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.224567 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.224608 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.224619 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.224639 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.224651 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:50Z","lastTransitionTime":"2025-12-11T08:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.327661 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.327718 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.327731 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.327749 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.327762 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:50Z","lastTransitionTime":"2025-12-11T08:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.430515 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.430559 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.430572 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.430591 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.430604 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:50Z","lastTransitionTime":"2025-12-11T08:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.533994 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.534054 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.534067 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.534086 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.534102 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:50Z","lastTransitionTime":"2025-12-11T08:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.636864 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.636940 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.636953 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.636982 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.636998 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:50Z","lastTransitionTime":"2025-12-11T08:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.739998 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.740058 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.740070 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.740090 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.740103 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:50Z","lastTransitionTime":"2025-12-11T08:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.842762 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.842824 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.842840 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.842897 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.842910 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:50Z","lastTransitionTime":"2025-12-11T08:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.945247 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.945279 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.945288 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.945302 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:50 crc kubenswrapper[4629]: I1211 08:37:50.945310 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:50Z","lastTransitionTime":"2025-12-11T08:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.048349 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.048391 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.048402 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.048420 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.048433 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:51Z","lastTransitionTime":"2025-12-11T08:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.150336 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.150385 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.150406 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.150422 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.150432 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:51Z","lastTransitionTime":"2025-12-11T08:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.253269 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.253310 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.253320 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.253335 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.253345 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:51Z","lastTransitionTime":"2025-12-11T08:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.356390 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.356451 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.356474 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.356503 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.356525 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:51Z","lastTransitionTime":"2025-12-11T08:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.459218 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.459255 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.459268 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.459286 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.459299 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:51Z","lastTransitionTime":"2025-12-11T08:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.562925 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.562988 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.562998 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.563015 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.563042 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:51Z","lastTransitionTime":"2025-12-11T08:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.666069 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.666149 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.666164 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.666204 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.666216 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:51Z","lastTransitionTime":"2025-12-11T08:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.768585 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.768634 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.768646 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.768666 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.768679 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:51Z","lastTransitionTime":"2025-12-11T08:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.872110 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.872155 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.872164 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.872189 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.872201 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:51Z","lastTransitionTime":"2025-12-11T08:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.975294 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.975352 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.975362 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.975377 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:51 crc kubenswrapper[4629]: I1211 08:37:51.975404 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:51Z","lastTransitionTime":"2025-12-11T08:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.078765 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.078798 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.078808 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.078822 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.078832 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:52Z","lastTransitionTime":"2025-12-11T08:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.181620 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.181667 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.181676 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.181693 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.181703 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:52Z","lastTransitionTime":"2025-12-11T08:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.198134 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:52 crc kubenswrapper[4629]: E1211 08:37:52.199039 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.199061 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.199208 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.199323 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:52 crc kubenswrapper[4629]: E1211 08:37:52.199366 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:52 crc kubenswrapper[4629]: E1211 08:37:52.199435 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:52 crc kubenswrapper[4629]: E1211 08:37:52.199493 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.210315 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.224098 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.244444 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:40Z\\\",\\\"message\\\":\\\"93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:40.084266 6197 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:37:40.084173 6197 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 08:37:40.084328 6197 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nF1211 08:37:40.084345 6197 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.256569 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.270726 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.283716 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.283785 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.283797 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.283859 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.283879 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:52Z","lastTransitionTime":"2025-12-11T08:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.286896 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.300579 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.314086 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.353246 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.373661 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.387070 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.387103 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.387115 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.387129 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.387139 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:52Z","lastTransitionTime":"2025-12-11T08:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.394011 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.419317 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.437188 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.451105 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.464020 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.483060 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.489178 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.489214 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.489223 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.489238 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.489247 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:52Z","lastTransitionTime":"2025-12-11T08:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.498813 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.510732 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:52Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.591880 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.591919 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.591929 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.591944 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.591954 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:52Z","lastTransitionTime":"2025-12-11T08:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.695272 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.695323 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.695334 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.695351 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.695361 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:52Z","lastTransitionTime":"2025-12-11T08:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.798149 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.798195 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.798203 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.798218 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.798233 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:52Z","lastTransitionTime":"2025-12-11T08:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.900708 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.900746 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.900757 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.900772 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:52 crc kubenswrapper[4629]: I1211 08:37:52.900783 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:52Z","lastTransitionTime":"2025-12-11T08:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.003892 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.003975 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.003992 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.004011 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.004054 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:53Z","lastTransitionTime":"2025-12-11T08:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.105869 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.105910 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.105921 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.105936 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.105947 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:53Z","lastTransitionTime":"2025-12-11T08:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.208523 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.208573 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.208582 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.208599 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.208613 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:53Z","lastTransitionTime":"2025-12-11T08:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.311063 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.311117 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.311128 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.311145 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.311158 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:53Z","lastTransitionTime":"2025-12-11T08:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.413964 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.414008 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.414019 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.414036 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.414046 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:53Z","lastTransitionTime":"2025-12-11T08:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.516756 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.517051 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.517144 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.517237 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.517317 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:53Z","lastTransitionTime":"2025-12-11T08:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.620619 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.620681 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.620691 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.620711 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.620721 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:53Z","lastTransitionTime":"2025-12-11T08:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.724076 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.724128 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.724141 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.724160 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.724177 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:53Z","lastTransitionTime":"2025-12-11T08:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.827104 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.827147 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.827157 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.827172 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.827181 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:53Z","lastTransitionTime":"2025-12-11T08:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.929477 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.929508 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.929518 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.929532 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:53 crc kubenswrapper[4629]: I1211 08:37:53.929542 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:53Z","lastTransitionTime":"2025-12-11T08:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.032171 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.032218 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.032234 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.032254 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.032264 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.092732 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.092789 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.092801 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.092821 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.092834 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: E1211 08:37:54.105247 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.108894 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.108968 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.108982 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.109004 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.109017 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: E1211 08:37:54.121293 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.125878 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.125922 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.125934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.125955 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.125973 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: E1211 08:37:54.138429 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.142205 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.142276 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.142288 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.142306 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.142349 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: E1211 08:37:54.155229 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.158869 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.158912 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.158924 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.158943 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.158956 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: E1211 08:37:54.170566 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:37:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:37:54 crc kubenswrapper[4629]: E1211 08:37:54.170742 4629 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.172663 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.172711 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.172720 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.172785 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.172799 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.198269 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.198309 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.198293 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:54 crc kubenswrapper[4629]: E1211 08:37:54.198492 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.198581 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:54 crc kubenswrapper[4629]: E1211 08:37:54.198683 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:54 crc kubenswrapper[4629]: E1211 08:37:54.198762 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:54 crc kubenswrapper[4629]: E1211 08:37:54.198801 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.275593 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.275642 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.275651 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.275669 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.275684 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.378288 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.379299 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.379495 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.379680 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.379951 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.488114 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.489007 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.489028 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.489058 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.489078 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.591572 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.591623 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.591633 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.591650 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.591660 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.693814 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.693859 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.693868 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.693882 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.693892 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.796563 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.796600 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.796612 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.796629 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.796639 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.899725 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.899776 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.899785 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.899799 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:54 crc kubenswrapper[4629]: I1211 08:37:54.899810 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:54Z","lastTransitionTime":"2025-12-11T08:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.002020 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.002063 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.002075 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.002091 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.002103 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:55Z","lastTransitionTime":"2025-12-11T08:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.104633 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.104716 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.104728 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.104757 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.104772 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:55Z","lastTransitionTime":"2025-12-11T08:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.199642 4629 scope.go:117] "RemoveContainer" containerID="3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758" Dec 11 08:37:55 crc kubenswrapper[4629]: E1211 08:37:55.199952 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.207353 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.207387 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.207396 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.207410 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.207420 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:55Z","lastTransitionTime":"2025-12-11T08:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.310999 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.311045 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.311054 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.311073 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.311083 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:55Z","lastTransitionTime":"2025-12-11T08:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.413671 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.413715 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.413725 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.413743 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.413757 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:55Z","lastTransitionTime":"2025-12-11T08:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.516871 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.516908 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.516917 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.516934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.516944 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:55Z","lastTransitionTime":"2025-12-11T08:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.619753 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.619945 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.619971 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.619994 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.620038 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:55Z","lastTransitionTime":"2025-12-11T08:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.722819 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.722869 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.722879 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.722894 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.722903 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:55Z","lastTransitionTime":"2025-12-11T08:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.826165 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.826228 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.826237 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.826254 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.826266 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:55Z","lastTransitionTime":"2025-12-11T08:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.929036 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.929106 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.929119 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.929135 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:55 crc kubenswrapper[4629]: I1211 08:37:55.929146 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:55Z","lastTransitionTime":"2025-12-11T08:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.031613 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.031675 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.031690 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.031712 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.031721 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:56Z","lastTransitionTime":"2025-12-11T08:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.134493 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.134557 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.134572 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.134594 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.134611 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:56Z","lastTransitionTime":"2025-12-11T08:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.198124 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.198195 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.198195 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.198216 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:56 crc kubenswrapper[4629]: E1211 08:37:56.198369 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:56 crc kubenswrapper[4629]: E1211 08:37:56.198497 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:56 crc kubenswrapper[4629]: E1211 08:37:56.198634 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:56 crc kubenswrapper[4629]: E1211 08:37:56.198752 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.237199 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.237252 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.237271 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.237290 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.237303 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:56Z","lastTransitionTime":"2025-12-11T08:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.340880 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.340934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.340946 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.340964 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.341029 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:56Z","lastTransitionTime":"2025-12-11T08:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.443526 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.443573 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.443586 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.443603 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.443615 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:56Z","lastTransitionTime":"2025-12-11T08:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.546900 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.546947 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.546959 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.546977 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.546991 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:56Z","lastTransitionTime":"2025-12-11T08:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.649333 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.649363 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.649371 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.649385 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.649393 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:56Z","lastTransitionTime":"2025-12-11T08:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.751613 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.751639 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.751647 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.751661 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.751670 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:56Z","lastTransitionTime":"2025-12-11T08:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.854802 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.854855 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.854867 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.854884 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.854893 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:56Z","lastTransitionTime":"2025-12-11T08:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.957621 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.957685 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.957699 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.957718 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:56 crc kubenswrapper[4629]: I1211 08:37:56.957730 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:56Z","lastTransitionTime":"2025-12-11T08:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.060809 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.060868 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.060878 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.060892 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.060904 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:57Z","lastTransitionTime":"2025-12-11T08:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.163090 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.163145 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.163154 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.163169 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.163181 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:57Z","lastTransitionTime":"2025-12-11T08:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.265763 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.265814 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.265827 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.265867 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.265880 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:57Z","lastTransitionTime":"2025-12-11T08:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.369324 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.369374 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.369385 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.369403 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.369415 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:57Z","lastTransitionTime":"2025-12-11T08:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.472130 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.472178 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.472192 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.472210 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.472223 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:57Z","lastTransitionTime":"2025-12-11T08:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.575115 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.575151 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.575160 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.575187 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.575198 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:57Z","lastTransitionTime":"2025-12-11T08:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.677302 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.677349 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.677361 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.677383 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.677396 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:57Z","lastTransitionTime":"2025-12-11T08:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.780116 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.780167 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.780177 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.780191 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.780201 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:57Z","lastTransitionTime":"2025-12-11T08:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.882560 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.882603 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.882614 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.882631 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.882645 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:57Z","lastTransitionTime":"2025-12-11T08:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.984436 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.984493 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.984506 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.984524 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:57 crc kubenswrapper[4629]: I1211 08:37:57.984535 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:57Z","lastTransitionTime":"2025-12-11T08:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.017893 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:58 crc kubenswrapper[4629]: E1211 08:37:58.018016 4629 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:58 crc kubenswrapper[4629]: E1211 08:37:58.018067 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs podName:13e996ab-2163-44ad-9ecc-052a44f0da4c nodeName:}" failed. No retries permitted until 2025-12-11 08:38:30.018049781 +0000 UTC m=+98.083465389 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs") pod "network-metrics-daemon-29fct" (UID: "13e996ab-2163-44ad-9ecc-052a44f0da4c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.087203 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.087272 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.087283 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.087300 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.087312 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:58Z","lastTransitionTime":"2025-12-11T08:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.190342 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.190392 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.190403 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.190424 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.190436 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:58Z","lastTransitionTime":"2025-12-11T08:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.199273 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.199337 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.199373 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:37:58 crc kubenswrapper[4629]: E1211 08:37:58.199466 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:37:58 crc kubenswrapper[4629]: E1211 08:37:58.199532 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.199558 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:37:58 crc kubenswrapper[4629]: E1211 08:37:58.199730 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:37:58 crc kubenswrapper[4629]: E1211 08:37:58.199772 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.292736 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.292766 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.292778 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.292795 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.292806 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:58Z","lastTransitionTime":"2025-12-11T08:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.395616 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.395666 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.395677 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.395695 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.395705 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:58Z","lastTransitionTime":"2025-12-11T08:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.498631 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.498682 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.498695 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.498713 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.498726 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:58Z","lastTransitionTime":"2025-12-11T08:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.600978 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.601007 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.601015 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.601029 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.601219 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:58Z","lastTransitionTime":"2025-12-11T08:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.703881 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.703949 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.703967 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.703991 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.704009 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:58Z","lastTransitionTime":"2025-12-11T08:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.807006 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.807041 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.807052 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.807070 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.807082 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:58Z","lastTransitionTime":"2025-12-11T08:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.909732 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.909775 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.909786 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.909804 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:58 crc kubenswrapper[4629]: I1211 08:37:58.909816 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:58Z","lastTransitionTime":"2025-12-11T08:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.012949 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.012998 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.013007 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.013022 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.013032 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:59Z","lastTransitionTime":"2025-12-11T08:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.115466 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.115505 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.115515 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.115529 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.115541 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:59Z","lastTransitionTime":"2025-12-11T08:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.218051 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.218094 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.218107 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.218123 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.218161 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:59Z","lastTransitionTime":"2025-12-11T08:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.320480 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.320532 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.320544 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.320560 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.320571 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:59Z","lastTransitionTime":"2025-12-11T08:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.423181 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.423226 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.423237 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.423251 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.423260 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:59Z","lastTransitionTime":"2025-12-11T08:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.526158 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.526193 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.526203 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.526218 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.526227 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:59Z","lastTransitionTime":"2025-12-11T08:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.629155 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.629202 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.629213 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.629231 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.629244 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:59Z","lastTransitionTime":"2025-12-11T08:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.732520 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.732566 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.732575 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.732591 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.732601 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:59Z","lastTransitionTime":"2025-12-11T08:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.834739 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.834776 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.834784 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.834798 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.834807 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:59Z","lastTransitionTime":"2025-12-11T08:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.938472 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.938518 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.938529 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.938545 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:37:59 crc kubenswrapper[4629]: I1211 08:37:59.938557 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:37:59Z","lastTransitionTime":"2025-12-11T08:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.041621 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.041684 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.041694 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.041732 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.041744 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:00Z","lastTransitionTime":"2025-12-11T08:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.144822 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.144889 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.144905 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.144926 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.144937 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:00Z","lastTransitionTime":"2025-12-11T08:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.198905 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.198955 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.198965 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.198905 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:00 crc kubenswrapper[4629]: E1211 08:38:00.199103 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:00 crc kubenswrapper[4629]: E1211 08:38:00.199179 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:00 crc kubenswrapper[4629]: E1211 08:38:00.199327 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:00 crc kubenswrapper[4629]: E1211 08:38:00.199416 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.247793 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.247841 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.247869 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.247890 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.247902 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:00Z","lastTransitionTime":"2025-12-11T08:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.350921 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.351244 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.351254 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.351270 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.351279 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:00Z","lastTransitionTime":"2025-12-11T08:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.453129 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.453167 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.453177 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.453211 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.453223 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:00Z","lastTransitionTime":"2025-12-11T08:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.555261 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.555351 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.555368 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.555390 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.555399 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:00Z","lastTransitionTime":"2025-12-11T08:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.633714 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lvjmx_f5568b3d-2e9b-4038-bb16-4ad3d87435ee/kube-multus/0.log" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.633771 4629 generic.go:334] "Generic (PLEG): container finished" podID="f5568b3d-2e9b-4038-bb16-4ad3d87435ee" containerID="26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba" exitCode=1 Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.633820 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lvjmx" event={"ID":"f5568b3d-2e9b-4038-bb16-4ad3d87435ee","Type":"ContainerDied","Data":"26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba"} Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.634383 4629 scope.go:117] "RemoveContainer" containerID="26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.652354 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.660287 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.660333 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.660343 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.660360 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.660373 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:00Z","lastTransitionTime":"2025-12-11T08:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.671764 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.682025 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.702568 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.718511 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.731160 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.747129 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.757122 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.763229 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.763295 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.763306 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.763325 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.763336 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:00Z","lastTransitionTime":"2025-12-11T08:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.771959 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.785917 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.797223 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.807779 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.819023 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.830391 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.843964 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.854217 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.866760 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.866805 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.866814 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.866857 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.866872 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:00Z","lastTransitionTime":"2025-12-11T08:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.866909 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:38:00Z\\\",\\\"message\\\":\\\"2025-12-11T08:37:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6\\\\n2025-12-11T08:37:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6 to /host/opt/cni/bin/\\\\n2025-12-11T08:37:15Z [verbose] multus-daemon started\\\\n2025-12-11T08:37:15Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:38:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.886334 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:40Z\\\",\\\"message\\\":\\\"93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:40.084266 6197 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:37:40.084173 6197 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 08:37:40.084328 6197 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nF1211 08:37:40.084345 6197 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.969446 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.969491 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.969504 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.969523 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:00 crc kubenswrapper[4629]: I1211 08:38:00.969535 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:00Z","lastTransitionTime":"2025-12-11T08:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.072224 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.072277 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.072288 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.072307 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.072323 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:01Z","lastTransitionTime":"2025-12-11T08:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.174725 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.174777 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.174787 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.174802 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.174812 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:01Z","lastTransitionTime":"2025-12-11T08:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.277579 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.277623 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.277634 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.277651 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.277662 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:01Z","lastTransitionTime":"2025-12-11T08:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.380113 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.380159 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.380170 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.380187 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.380197 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:01Z","lastTransitionTime":"2025-12-11T08:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.482162 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.482204 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.482217 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.482235 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.482248 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:01Z","lastTransitionTime":"2025-12-11T08:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.584444 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.584484 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.584493 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.584510 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.584521 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:01Z","lastTransitionTime":"2025-12-11T08:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.637741 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lvjmx_f5568b3d-2e9b-4038-bb16-4ad3d87435ee/kube-multus/0.log" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.637796 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lvjmx" event={"ID":"f5568b3d-2e9b-4038-bb16-4ad3d87435ee","Type":"ContainerStarted","Data":"9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6"} Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.653753 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:38:00Z\\\",\\\"message\\\":\\\"2025-12-11T08:37:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6\\\\n2025-12-11T08:37:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6 to /host/opt/cni/bin/\\\\n2025-12-11T08:37:15Z [verbose] multus-daemon started\\\\n2025-12-11T08:37:15Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:38:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.673744 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:40Z\\\",\\\"message\\\":\\\"93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:40.084266 6197 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:37:40.084173 6197 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 08:37:40.084328 6197 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nF1211 08:37:40.084345 6197 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.684060 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.686745 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.686789 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.686802 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.686819 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.686832 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:01Z","lastTransitionTime":"2025-12-11T08:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.696778 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.709956 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.721730 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.731531 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.742680 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.756689 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.767385 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.790092 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.790130 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.790142 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.790160 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.790172 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:01Z","lastTransitionTime":"2025-12-11T08:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.792026 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.806050 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.822729 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.836727 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.846589 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.858879 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.870414 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.882709 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.892501 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.892553 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.892565 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.892580 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.892591 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:01Z","lastTransitionTime":"2025-12-11T08:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.994826 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.994882 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.994895 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.994912 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:01 crc kubenswrapper[4629]: I1211 08:38:01.994923 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:01Z","lastTransitionTime":"2025-12-11T08:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.097740 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.097789 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.097800 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.097818 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.097832 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:02Z","lastTransitionTime":"2025-12-11T08:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.198568 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.198598 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.198606 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:02 crc kubenswrapper[4629]: E1211 08:38:02.198717 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.198751 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:02 crc kubenswrapper[4629]: E1211 08:38:02.198870 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:02 crc kubenswrapper[4629]: E1211 08:38:02.198972 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:02 crc kubenswrapper[4629]: E1211 08:38:02.199019 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.200225 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.200247 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.200255 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.200267 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.200275 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:02Z","lastTransitionTime":"2025-12-11T08:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.214287 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.226603 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.245653 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.257217 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.269140 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.282562 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.296444 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.302062 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.302109 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.302123 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.302140 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.302152 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:02Z","lastTransitionTime":"2025-12-11T08:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.308832 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.322351 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.334268 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.346566 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:38:00Z\\\",\\\"message\\\":\\\"2025-12-11T08:37:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6\\\\n2025-12-11T08:37:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6 to /host/opt/cni/bin/\\\\n2025-12-11T08:37:15Z [verbose] multus-daemon started\\\\n2025-12-11T08:37:15Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:38:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.365046 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:40Z\\\",\\\"message\\\":\\\"93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:40.084266 6197 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:37:40.084173 6197 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 08:37:40.084328 6197 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nF1211 08:37:40.084345 6197 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.374888 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.386706 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.399608 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.404030 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.404062 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.404070 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.404085 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.404095 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:02Z","lastTransitionTime":"2025-12-11T08:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.410710 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.424322 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.436460 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:02Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.506439 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.506480 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.506491 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.506551 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.506565 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:02Z","lastTransitionTime":"2025-12-11T08:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.609029 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.609068 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.609077 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.609094 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.609103 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:02Z","lastTransitionTime":"2025-12-11T08:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.711824 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.711891 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.711901 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.711920 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.711933 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:02Z","lastTransitionTime":"2025-12-11T08:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.814027 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.814299 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.814403 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.814500 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.814589 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:02Z","lastTransitionTime":"2025-12-11T08:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.916909 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.916948 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.916959 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.916977 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:02 crc kubenswrapper[4629]: I1211 08:38:02.916988 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:02Z","lastTransitionTime":"2025-12-11T08:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.019110 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.019474 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.019609 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.019712 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.019822 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:03Z","lastTransitionTime":"2025-12-11T08:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.121834 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.122177 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.122291 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.122388 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.122479 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:03Z","lastTransitionTime":"2025-12-11T08:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.225008 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.225288 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.225373 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.225456 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.225521 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:03Z","lastTransitionTime":"2025-12-11T08:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.328448 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.328505 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.328515 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.328533 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.328542 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:03Z","lastTransitionTime":"2025-12-11T08:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.430718 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.430756 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.430766 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.430780 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.430789 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:03Z","lastTransitionTime":"2025-12-11T08:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.533551 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.533587 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.533597 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.533612 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.533621 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:03Z","lastTransitionTime":"2025-12-11T08:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.636432 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.636478 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.636490 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.636508 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.636520 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:03Z","lastTransitionTime":"2025-12-11T08:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.738688 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.738725 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.738735 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.738755 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.738773 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:03Z","lastTransitionTime":"2025-12-11T08:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.841248 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.841488 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.841591 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.841677 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.841760 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:03Z","lastTransitionTime":"2025-12-11T08:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.944398 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.944441 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.944452 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.944470 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:03 crc kubenswrapper[4629]: I1211 08:38:03.944479 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:03Z","lastTransitionTime":"2025-12-11T08:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.046583 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.046797 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.046895 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.046974 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.047043 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.149578 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.149619 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.149627 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.149642 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.149652 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.198540 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.198568 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.198603 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.198752 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:04 crc kubenswrapper[4629]: E1211 08:38:04.198910 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:04 crc kubenswrapper[4629]: E1211 08:38:04.199100 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:04 crc kubenswrapper[4629]: E1211 08:38:04.199203 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:04 crc kubenswrapper[4629]: E1211 08:38:04.199284 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.251726 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.251765 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.251774 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.251789 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.251799 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.354666 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.354728 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.354741 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.354759 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.354771 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.457044 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.457094 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.457109 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.457129 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.457145 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.495925 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.495961 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.495969 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.495983 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.495995 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: E1211 08:38:04.508062 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:04Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.511880 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.511919 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.511928 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.511946 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.511960 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: E1211 08:38:04.523002 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:04Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.525793 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.525827 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.525835 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.525861 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.525871 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: E1211 08:38:04.536326 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:04Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.539017 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.539057 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.539066 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.539082 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.539092 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: E1211 08:38:04.549174 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:04Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.553216 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.553259 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.553271 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.553288 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.553300 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: E1211 08:38:04.568583 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:04Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:04 crc kubenswrapper[4629]: E1211 08:38:04.569004 4629 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.571150 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.571199 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.571212 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.571229 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.571241 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.674189 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.674236 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.674245 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.674261 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.674270 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.776656 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.776980 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.777086 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.777189 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.777284 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.880044 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.880115 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.880128 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.880147 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.880158 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.982593 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.982637 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.982648 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.982667 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:04 crc kubenswrapper[4629]: I1211 08:38:04.982680 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:04Z","lastTransitionTime":"2025-12-11T08:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.084724 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.084765 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.084776 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.084796 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.084807 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:05Z","lastTransitionTime":"2025-12-11T08:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.187188 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.187228 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.187241 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.187257 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.187269 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:05Z","lastTransitionTime":"2025-12-11T08:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.289534 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.289579 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.289587 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.289601 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.289610 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:05Z","lastTransitionTime":"2025-12-11T08:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.392116 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.392161 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.392179 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.392202 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.392219 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:05Z","lastTransitionTime":"2025-12-11T08:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.493825 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.493905 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.493921 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.493970 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.493987 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:05Z","lastTransitionTime":"2025-12-11T08:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.595973 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.596031 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.596046 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.596063 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.596074 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:05Z","lastTransitionTime":"2025-12-11T08:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.698359 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.698419 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.698429 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.698446 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.698460 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:05Z","lastTransitionTime":"2025-12-11T08:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.800930 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.800986 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.800995 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.801010 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.801019 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:05Z","lastTransitionTime":"2025-12-11T08:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.904087 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.904138 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.904148 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.904162 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:05 crc kubenswrapper[4629]: I1211 08:38:05.904171 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:05Z","lastTransitionTime":"2025-12-11T08:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.006870 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.006907 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.006916 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.006934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.006943 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:06Z","lastTransitionTime":"2025-12-11T08:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.109041 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.109090 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.109101 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.109120 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.109131 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:06Z","lastTransitionTime":"2025-12-11T08:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.198006 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.198083 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.198137 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.198095 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:06 crc kubenswrapper[4629]: E1211 08:38:06.198227 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:06 crc kubenswrapper[4629]: E1211 08:38:06.198325 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:06 crc kubenswrapper[4629]: E1211 08:38:06.198391 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:06 crc kubenswrapper[4629]: E1211 08:38:06.198458 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.211268 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.211299 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.211308 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.211320 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.211329 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:06Z","lastTransitionTime":"2025-12-11T08:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.314205 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.314237 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.314245 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.314261 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.314271 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:06Z","lastTransitionTime":"2025-12-11T08:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.417491 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.417525 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.417533 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.417547 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.417556 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:06Z","lastTransitionTime":"2025-12-11T08:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.521986 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.522045 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.522059 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.522080 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.522092 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:06Z","lastTransitionTime":"2025-12-11T08:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.624918 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.624963 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.624976 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.624994 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.625004 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:06Z","lastTransitionTime":"2025-12-11T08:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.727272 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.727331 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.727341 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.727377 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.727388 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:06Z","lastTransitionTime":"2025-12-11T08:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.830201 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.830234 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.830242 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.830256 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.830266 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:06Z","lastTransitionTime":"2025-12-11T08:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.933090 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.933391 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.933475 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.933553 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:06 crc kubenswrapper[4629]: I1211 08:38:06.933630 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:06Z","lastTransitionTime":"2025-12-11T08:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.036266 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.036303 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.036312 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.036326 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.036338 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:07Z","lastTransitionTime":"2025-12-11T08:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.138749 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.138794 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.138802 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.138817 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.138826 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:07Z","lastTransitionTime":"2025-12-11T08:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.199378 4629 scope.go:117] "RemoveContainer" containerID="3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.242202 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.242236 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.242247 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.242264 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.242276 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:07Z","lastTransitionTime":"2025-12-11T08:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.345559 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.345599 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.345608 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.345625 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.345634 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:07Z","lastTransitionTime":"2025-12-11T08:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.448426 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.448475 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.448485 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.448502 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.448511 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:07Z","lastTransitionTime":"2025-12-11T08:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.551731 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.551775 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.551789 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.551808 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.551823 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:07Z","lastTransitionTime":"2025-12-11T08:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.653810 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.653868 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.653878 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.653896 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.653906 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:07Z","lastTransitionTime":"2025-12-11T08:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.656756 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/2.log" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.659645 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9"} Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.660460 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.673819 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.687434 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.702986 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.718223 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.732623 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.745145 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.757687 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.758413 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.758451 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.758461 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.758479 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.758492 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:07Z","lastTransitionTime":"2025-12-11T08:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.774003 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:38:00Z\\\",\\\"message\\\":\\\"2025-12-11T08:37:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6\\\\n2025-12-11T08:37:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6 to /host/opt/cni/bin/\\\\n2025-12-11T08:37:15Z [verbose] multus-daemon started\\\\n2025-12-11T08:37:15Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:38:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.802297 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:40Z\\\",\\\"message\\\":\\\"93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:40.084266 6197 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:37:40.084173 6197 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 08:37:40.084328 6197 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nF1211 08:37:40.084345 6197 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:38:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.818324 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.831636 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.852041 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.861389 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.861453 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.861464 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.861483 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.861496 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:07Z","lastTransitionTime":"2025-12-11T08:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.871199 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.897163 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.911809 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.927804 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.942266 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.952314 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.964927 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.964976 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.964987 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.965009 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:07 crc kubenswrapper[4629]: I1211 08:38:07.965023 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:07Z","lastTransitionTime":"2025-12-11T08:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.067040 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.067097 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.067106 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.067122 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.067132 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:08Z","lastTransitionTime":"2025-12-11T08:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.169795 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.169876 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.169886 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.169916 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.169953 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:08Z","lastTransitionTime":"2025-12-11T08:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.198140 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.198234 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:08 crc kubenswrapper[4629]: E1211 08:38:08.198282 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.198249 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:08 crc kubenswrapper[4629]: E1211 08:38:08.198457 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:08 crc kubenswrapper[4629]: E1211 08:38:08.198560 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.198582 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:08 crc kubenswrapper[4629]: E1211 08:38:08.198639 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.273369 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.273447 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.273462 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.273482 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.273499 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:08Z","lastTransitionTime":"2025-12-11T08:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.376783 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.376832 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.376841 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.376870 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.376881 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:08Z","lastTransitionTime":"2025-12-11T08:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.479053 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.479099 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.479109 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.479125 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.479137 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:08Z","lastTransitionTime":"2025-12-11T08:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.581627 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.581677 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.581720 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.581738 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.581749 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:08Z","lastTransitionTime":"2025-12-11T08:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.665475 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/3.log" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.666575 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/2.log" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.669185 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" exitCode=1 Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.669256 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9"} Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.669327 4629 scope.go:117] "RemoveContainer" containerID="3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.670156 4629 scope.go:117] "RemoveContainer" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" Dec 11 08:38:08 crc kubenswrapper[4629]: E1211 08:38:08.670402 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.685669 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.686157 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.686190 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.686199 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.686213 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.686224 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:08Z","lastTransitionTime":"2025-12-11T08:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.701448 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.714638 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.734663 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.748998 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.762242 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.776512 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.785824 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.789697 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.789730 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.789738 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.789752 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.789761 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:08Z","lastTransitionTime":"2025-12-11T08:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.799196 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.813264 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.825468 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.844548 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:38:00Z\\\",\\\"message\\\":\\\"2025-12-11T08:37:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6\\\\n2025-12-11T08:37:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6 to /host/opt/cni/bin/\\\\n2025-12-11T08:37:15Z [verbose] multus-daemon started\\\\n2025-12-11T08:37:15Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:38:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.866790 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f18adc233822b5eed12cb34bd1f4d046316b3468b3383c429a921d7283f8758\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:37:40Z\\\",\\\"message\\\":\\\"93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1211 08:37:40.084266 6197 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:37:40.084173 6197 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 08:37:40.084328 6197 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nF1211 08:37:40.084345 6197 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:38:08Z\\\",\\\"message\\\":\\\"openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z]\\\\nI1211 08:38:08.231297 6582 kube.go:317] Updating pod openshift-multus/network-metrics-daemon-29fct\\\\nI1211 08:38:08.231275 6582 services_controller.go:434] Service openshift-dns-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-dns-operator 4bf7a6e2-037e-4e09-ad6b-2e7f1059a532 4106 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:dns-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc000701b47 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: dns-ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:38:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.878617 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.889913 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.891966 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.892000 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.892009 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.892038 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.892047 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:08Z","lastTransitionTime":"2025-12-11T08:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.902244 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.913928 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.924799 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.995020 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.995056 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.995067 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.995084 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:08 crc kubenswrapper[4629]: I1211 08:38:08.995096 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:08Z","lastTransitionTime":"2025-12-11T08:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.097562 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.097625 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.097636 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.097650 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.097659 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:09Z","lastTransitionTime":"2025-12-11T08:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.199961 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.200007 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.200020 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.200036 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.200046 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:09Z","lastTransitionTime":"2025-12-11T08:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.302561 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.302597 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.302608 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.302624 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.302634 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:09Z","lastTransitionTime":"2025-12-11T08:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.405238 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.405284 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.405292 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.405308 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.405317 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:09Z","lastTransitionTime":"2025-12-11T08:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.508170 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.508241 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.508257 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.508275 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.508287 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:09Z","lastTransitionTime":"2025-12-11T08:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.611263 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.611339 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.611354 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.611374 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.611389 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:09Z","lastTransitionTime":"2025-12-11T08:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.673358 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/3.log" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.679107 4629 scope.go:117] "RemoveContainer" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" Dec 11 08:38:09 crc kubenswrapper[4629]: E1211 08:38:09.679307 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.692264 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.706613 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:38:00Z\\\",\\\"message\\\":\\\"2025-12-11T08:37:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6\\\\n2025-12-11T08:37:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6 to /host/opt/cni/bin/\\\\n2025-12-11T08:37:15Z [verbose] multus-daemon started\\\\n2025-12-11T08:37:15Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:38:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.713990 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.714046 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.714059 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.714079 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.714092 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:09Z","lastTransitionTime":"2025-12-11T08:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.726390 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:38:08Z\\\",\\\"message\\\":\\\"openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z]\\\\nI1211 08:38:08.231297 6582 kube.go:317] Updating pod openshift-multus/network-metrics-daemon-29fct\\\\nI1211 08:38:08.231275 6582 services_controller.go:434] Service openshift-dns-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-dns-operator 4bf7a6e2-037e-4e09-ad6b-2e7f1059a532 4106 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:dns-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc000701b47 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: dns-ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:38:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.737607 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.749540 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.761829 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.776710 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.791968 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.805307 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.816368 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.816418 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.816430 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.816449 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.816461 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:09Z","lastTransitionTime":"2025-12-11T08:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.816970 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.825602 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.843268 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.858459 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.870440 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.880817 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.902422 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.919136 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.919190 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.919201 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.919222 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.919234 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:09Z","lastTransitionTime":"2025-12-11T08:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.935802 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:09 crc kubenswrapper[4629]: I1211 08:38:09.960971 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:09Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.021959 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.022001 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.022013 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.022033 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.022044 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:10Z","lastTransitionTime":"2025-12-11T08:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.124331 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.124603 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.124687 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.124775 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.124838 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:10Z","lastTransitionTime":"2025-12-11T08:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.198266 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.198296 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.198271 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:10 crc kubenswrapper[4629]: E1211 08:38:10.198470 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.198573 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:10 crc kubenswrapper[4629]: E1211 08:38:10.198622 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:10 crc kubenswrapper[4629]: E1211 08:38:10.198662 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:10 crc kubenswrapper[4629]: E1211 08:38:10.198695 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.213399 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.227923 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.227955 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.227963 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.227978 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.227988 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:10Z","lastTransitionTime":"2025-12-11T08:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.329963 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.330001 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.330012 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.330029 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.330041 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:10Z","lastTransitionTime":"2025-12-11T08:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.432585 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.432625 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.432635 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.432651 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.432662 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:10Z","lastTransitionTime":"2025-12-11T08:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.536342 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.536409 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.536427 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.536452 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.536470 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:10Z","lastTransitionTime":"2025-12-11T08:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.639129 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.639197 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.639215 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.639240 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.639260 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:10Z","lastTransitionTime":"2025-12-11T08:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.742716 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.742757 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.742767 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.742784 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.742794 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:10Z","lastTransitionTime":"2025-12-11T08:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.846272 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.846830 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.848002 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.848057 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.848071 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:10Z","lastTransitionTime":"2025-12-11T08:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.951081 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.951121 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.951133 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.951147 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:10 crc kubenswrapper[4629]: I1211 08:38:10.951159 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:10Z","lastTransitionTime":"2025-12-11T08:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.053857 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.054168 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.054237 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.054307 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.054379 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:11Z","lastTransitionTime":"2025-12-11T08:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.156519 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.156558 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.156569 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.156584 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.156594 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:11Z","lastTransitionTime":"2025-12-11T08:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.258743 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.259090 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.259181 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.259274 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.259370 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:11Z","lastTransitionTime":"2025-12-11T08:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.361786 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.361837 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.361879 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.361897 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.361907 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:11Z","lastTransitionTime":"2025-12-11T08:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.465188 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.465626 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.465779 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.465983 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.466346 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:11Z","lastTransitionTime":"2025-12-11T08:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.568887 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.568934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.569047 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.569085 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.569110 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:11Z","lastTransitionTime":"2025-12-11T08:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.673349 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.673993 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.674076 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.674174 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.674247 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:11Z","lastTransitionTime":"2025-12-11T08:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.777571 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.777611 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.777622 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.777638 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.777651 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:11Z","lastTransitionTime":"2025-12-11T08:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.880038 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.880083 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.880094 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.880111 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.880126 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:11Z","lastTransitionTime":"2025-12-11T08:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.984520 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.984570 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.984581 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.984610 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:11 crc kubenswrapper[4629]: I1211 08:38:11.984635 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:11Z","lastTransitionTime":"2025-12-11T08:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.087423 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.087746 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.087874 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.087998 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.088091 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:12Z","lastTransitionTime":"2025-12-11T08:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.192191 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.192573 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.192906 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.193139 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.193344 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:12Z","lastTransitionTime":"2025-12-11T08:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.199091 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.199123 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.199486 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:12 crc kubenswrapper[4629]: E1211 08:38:12.199484 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:12 crc kubenswrapper[4629]: E1211 08:38:12.199541 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:12 crc kubenswrapper[4629]: E1211 08:38:12.199763 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.200058 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:12 crc kubenswrapper[4629]: E1211 08:38:12.200215 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.214518 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.225700 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.236454 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.247000 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.260790 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:38:00Z\\\",\\\"message\\\":\\\"2025-12-11T08:37:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6\\\\n2025-12-11T08:37:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6 to /host/opt/cni/bin/\\\\n2025-12-11T08:37:15Z [verbose] multus-daemon started\\\\n2025-12-11T08:37:15Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:38:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.316256 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.316284 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.316292 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.316305 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.316314 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:12Z","lastTransitionTime":"2025-12-11T08:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.330860 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:38:08Z\\\",\\\"message\\\":\\\"openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z]\\\\nI1211 08:38:08.231297 6582 kube.go:317] Updating pod openshift-multus/network-metrics-daemon-29fct\\\\nI1211 08:38:08.231275 6582 services_controller.go:434] Service openshift-dns-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-dns-operator 4bf7a6e2-037e-4e09-ad6b-2e7f1059a532 4106 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:dns-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc000701b47 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: dns-ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:38:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.342339 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.352026 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0e625b2-b7fd-4aed-a513-8ad7f8c5bd52\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8edcf812544fff27a0105abbaae88e57c68b9823483aa578a21e79ccd342c8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://763ac579712e110e6989f12f4e7ab5ed267017bae7b82f3a28b0af7a92274bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://763ac579712e110e6989f12f4e7ab5ed267017bae7b82f3a28b0af7a92274bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.362736 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.374153 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.390794 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.403518 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.418735 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.418956 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.419043 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.418659 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.419131 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.419453 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:12Z","lastTransitionTime":"2025-12-11T08:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.483208 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.494085 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.513585 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.522124 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.522158 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.522167 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.522182 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.522191 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:12Z","lastTransitionTime":"2025-12-11T08:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.529472 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.544534 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.562915 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:12Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.624791 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.624826 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.624835 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.624864 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.624875 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:12Z","lastTransitionTime":"2025-12-11T08:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.727105 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.727150 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.727163 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.727182 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.727194 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:12Z","lastTransitionTime":"2025-12-11T08:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.832420 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.832463 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.832473 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.832491 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.832502 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:12Z","lastTransitionTime":"2025-12-11T08:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.935613 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.935673 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.935692 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.935719 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:12 crc kubenswrapper[4629]: I1211 08:38:12.935749 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:12Z","lastTransitionTime":"2025-12-11T08:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.038980 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.039023 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.039033 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.039049 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.039062 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:13Z","lastTransitionTime":"2025-12-11T08:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.142302 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.142366 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.142419 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.142445 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.142458 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:13Z","lastTransitionTime":"2025-12-11T08:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.244358 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.244534 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.244553 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.244570 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.244582 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:13Z","lastTransitionTime":"2025-12-11T08:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.346434 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.346467 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.346476 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.346489 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.346497 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:13Z","lastTransitionTime":"2025-12-11T08:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.448527 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.448565 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.448575 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.448591 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.448601 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:13Z","lastTransitionTime":"2025-12-11T08:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.550160 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.550199 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.550209 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.550223 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.550235 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:13Z","lastTransitionTime":"2025-12-11T08:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.653104 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.653155 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.653167 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.653186 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.653198 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:13Z","lastTransitionTime":"2025-12-11T08:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.759492 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.759549 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.759566 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.759583 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.759594 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:13Z","lastTransitionTime":"2025-12-11T08:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.862085 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.862132 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.862144 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.862162 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.862182 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:13Z","lastTransitionTime":"2025-12-11T08:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.964785 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.964866 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.964886 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.964905 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:13 crc kubenswrapper[4629]: I1211 08:38:13.964916 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:13Z","lastTransitionTime":"2025-12-11T08:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.068041 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.068081 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.068091 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.068108 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.068121 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.135214 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.135374 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.135416 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.135448 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.135663 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.135744 4629 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.135909 4629 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.135918 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.135902826 +0000 UTC m=+146.201318434 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.135982 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.136010 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.136014 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.135993329 +0000 UTC m=+146.201409037 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.136031 4629 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.136035 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.136065 4629 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.136085 4629 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.136067 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.136058341 +0000 UTC m=+146.201473949 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.136188 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.136165875 +0000 UTC m=+146.201581553 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.136210 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.136199876 +0000 UTC m=+146.201615614 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.170995 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.171034 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.171044 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.171058 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.171067 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.198298 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.198338 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.198349 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.198419 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.198426 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.198534 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.198597 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.198662 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.273724 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.273765 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.273773 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.273789 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.273799 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.377154 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.377195 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.377204 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.377217 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.377226 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.479873 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.479905 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.479914 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.479929 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.479938 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.582191 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.582243 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.582253 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.582267 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.582277 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.583472 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.583525 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.583563 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.583579 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.583588 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.599823 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.604193 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.604271 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.604286 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.604300 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.604309 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.617592 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.621674 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.621738 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.621754 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.621773 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.621785 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.637103 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.643500 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.643568 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.643587 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.643605 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.643615 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.656152 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.659976 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.660032 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.660043 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.660059 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.660071 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.672912 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:14Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:14 crc kubenswrapper[4629]: E1211 08:38:14.673088 4629 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.684893 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.684932 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.684941 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.684954 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.684963 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.788069 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.788115 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.788125 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.788143 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.788153 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.890656 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.890723 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.890738 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.890754 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.890765 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.993382 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.993420 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.993428 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.993440 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:14 crc kubenswrapper[4629]: I1211 08:38:14.993448 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:14Z","lastTransitionTime":"2025-12-11T08:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.095474 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.095535 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.095546 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.095564 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.095575 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:15Z","lastTransitionTime":"2025-12-11T08:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.198182 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.198239 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.198251 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.198267 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.198281 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:15Z","lastTransitionTime":"2025-12-11T08:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.301025 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.301069 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.301079 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.301094 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.301102 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:15Z","lastTransitionTime":"2025-12-11T08:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.403433 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.403471 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.403479 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.403493 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.403503 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:15Z","lastTransitionTime":"2025-12-11T08:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.505635 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.505683 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.505697 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.505714 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.505725 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:15Z","lastTransitionTime":"2025-12-11T08:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.608351 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.608394 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.608403 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.608418 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.608428 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:15Z","lastTransitionTime":"2025-12-11T08:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.710790 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.710841 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.710868 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.710884 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.710918 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:15Z","lastTransitionTime":"2025-12-11T08:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.814225 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.814273 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.814282 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.814297 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.814308 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:15Z","lastTransitionTime":"2025-12-11T08:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.917806 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.917896 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.917914 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.917933 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:15 crc kubenswrapper[4629]: I1211 08:38:15.917947 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:15Z","lastTransitionTime":"2025-12-11T08:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.019861 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.019913 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.019925 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.019942 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.019954 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:16Z","lastTransitionTime":"2025-12-11T08:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.122471 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.122523 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.122534 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.122549 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.122559 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:16Z","lastTransitionTime":"2025-12-11T08:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.198535 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.198596 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:16 crc kubenswrapper[4629]: E1211 08:38:16.198682 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:16 crc kubenswrapper[4629]: E1211 08:38:16.198757 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.198890 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:16 crc kubenswrapper[4629]: E1211 08:38:16.198957 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.199045 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:16 crc kubenswrapper[4629]: E1211 08:38:16.199122 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.225036 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.225086 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.225098 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.225117 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.225129 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:16Z","lastTransitionTime":"2025-12-11T08:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.327547 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.327621 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.327633 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.327650 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.327663 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:16Z","lastTransitionTime":"2025-12-11T08:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.430331 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.430376 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.430392 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.430407 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.430421 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:16Z","lastTransitionTime":"2025-12-11T08:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.533510 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.533556 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.533565 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.533580 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.533597 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:16Z","lastTransitionTime":"2025-12-11T08:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.635714 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.635760 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.635769 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.635783 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.635796 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:16Z","lastTransitionTime":"2025-12-11T08:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.738197 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.738249 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.738260 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.738276 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.738288 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:16Z","lastTransitionTime":"2025-12-11T08:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.840289 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.840364 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.840375 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.840391 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.840400 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:16Z","lastTransitionTime":"2025-12-11T08:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.942977 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.943013 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.943022 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.943038 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:16 crc kubenswrapper[4629]: I1211 08:38:16.943048 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:16Z","lastTransitionTime":"2025-12-11T08:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.046445 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.046487 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.046496 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.046510 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.046519 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:17Z","lastTransitionTime":"2025-12-11T08:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.149291 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.149323 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.149331 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.149344 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.149353 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:17Z","lastTransitionTime":"2025-12-11T08:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.251641 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.251673 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.251682 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.251696 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.251707 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:17Z","lastTransitionTime":"2025-12-11T08:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.354698 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.354735 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.354749 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.354764 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.354774 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:17Z","lastTransitionTime":"2025-12-11T08:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.457692 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.457733 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.457745 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.457761 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.457770 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:17Z","lastTransitionTime":"2025-12-11T08:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.560249 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.560303 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.560311 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.560325 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.560336 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:17Z","lastTransitionTime":"2025-12-11T08:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.662992 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.663069 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.663079 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.663096 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.663106 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:17Z","lastTransitionTime":"2025-12-11T08:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.765704 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.765745 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.765757 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.765771 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.765783 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:17Z","lastTransitionTime":"2025-12-11T08:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.868064 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.868097 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.868106 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.868119 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.868128 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:17Z","lastTransitionTime":"2025-12-11T08:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.971048 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.971389 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.971399 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.971415 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:17 crc kubenswrapper[4629]: I1211 08:38:17.971424 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:17Z","lastTransitionTime":"2025-12-11T08:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.074134 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.074176 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.074188 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.074204 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.074216 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:18Z","lastTransitionTime":"2025-12-11T08:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.176984 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.177040 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.177050 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.177064 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.177073 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:18Z","lastTransitionTime":"2025-12-11T08:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.198792 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.198832 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.198906 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.198797 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:18 crc kubenswrapper[4629]: E1211 08:38:18.198993 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:18 crc kubenswrapper[4629]: E1211 08:38:18.199211 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:18 crc kubenswrapper[4629]: E1211 08:38:18.199211 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:18 crc kubenswrapper[4629]: E1211 08:38:18.199277 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.280425 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.280475 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.280490 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.280509 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.280523 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:18Z","lastTransitionTime":"2025-12-11T08:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.383461 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.383530 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.383543 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.383560 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.383574 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:18Z","lastTransitionTime":"2025-12-11T08:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.485911 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.485975 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.485995 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.486016 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.486030 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:18Z","lastTransitionTime":"2025-12-11T08:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.588402 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.588455 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.588465 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.588479 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.588490 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:18Z","lastTransitionTime":"2025-12-11T08:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.692496 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.692568 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.692582 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.692602 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.692614 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:18Z","lastTransitionTime":"2025-12-11T08:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.795736 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.795788 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.795797 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.795811 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.795820 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:18Z","lastTransitionTime":"2025-12-11T08:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.898175 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.898254 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.898267 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.898287 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:18 crc kubenswrapper[4629]: I1211 08:38:18.898306 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:18Z","lastTransitionTime":"2025-12-11T08:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.000490 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.000547 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.000560 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.000579 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.000591 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:19Z","lastTransitionTime":"2025-12-11T08:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.103372 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.103420 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.103429 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.103444 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.103456 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:19Z","lastTransitionTime":"2025-12-11T08:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.206186 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.206219 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.206229 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.206241 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.206251 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:19Z","lastTransitionTime":"2025-12-11T08:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.308671 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.308705 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.308714 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.308730 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.308740 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:19Z","lastTransitionTime":"2025-12-11T08:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.411705 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.411770 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.411780 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.411795 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.411808 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:19Z","lastTransitionTime":"2025-12-11T08:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.514551 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.514593 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.514604 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.514635 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.514651 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:19Z","lastTransitionTime":"2025-12-11T08:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.618027 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.618161 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.618172 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.618189 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.618198 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:19Z","lastTransitionTime":"2025-12-11T08:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.720318 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.720372 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.720395 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.720415 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.720430 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:19Z","lastTransitionTime":"2025-12-11T08:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.822395 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.822437 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.822456 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.822490 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.822500 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:19Z","lastTransitionTime":"2025-12-11T08:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.924697 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.924761 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.924773 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.924788 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:19 crc kubenswrapper[4629]: I1211 08:38:19.924798 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:19Z","lastTransitionTime":"2025-12-11T08:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.027553 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.027596 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.027606 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.027621 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.027632 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:20Z","lastTransitionTime":"2025-12-11T08:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.130165 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.130211 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.130225 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.130241 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.130253 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:20Z","lastTransitionTime":"2025-12-11T08:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.198205 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.198273 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.198390 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:20 crc kubenswrapper[4629]: E1211 08:38:20.198496 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.198902 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:20 crc kubenswrapper[4629]: E1211 08:38:20.199038 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:20 crc kubenswrapper[4629]: E1211 08:38:20.199348 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:20 crc kubenswrapper[4629]: E1211 08:38:20.199630 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.232277 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.232320 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.232346 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.232364 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.232374 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:20Z","lastTransitionTime":"2025-12-11T08:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.334806 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.334885 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.334899 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.334918 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.334930 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:20Z","lastTransitionTime":"2025-12-11T08:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.437523 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.437563 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.437575 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.437591 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.437603 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:20Z","lastTransitionTime":"2025-12-11T08:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.540292 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.540333 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.540344 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.540362 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.540410 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:20Z","lastTransitionTime":"2025-12-11T08:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.642318 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.642390 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.642404 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.642422 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.642435 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:20Z","lastTransitionTime":"2025-12-11T08:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.746118 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.746167 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.746185 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.746202 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.746212 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:20Z","lastTransitionTime":"2025-12-11T08:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.848466 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.848521 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.848533 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.848553 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.848567 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:20Z","lastTransitionTime":"2025-12-11T08:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.950713 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.950759 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.950774 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.950792 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:20 crc kubenswrapper[4629]: I1211 08:38:20.950805 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:20Z","lastTransitionTime":"2025-12-11T08:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.053133 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.053172 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.053181 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.053195 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.053204 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:21Z","lastTransitionTime":"2025-12-11T08:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.155148 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.155184 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.155195 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.155211 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.155221 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:21Z","lastTransitionTime":"2025-12-11T08:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.200011 4629 scope.go:117] "RemoveContainer" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" Dec 11 08:38:21 crc kubenswrapper[4629]: E1211 08:38:21.200838 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.257962 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.258037 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.258055 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.258086 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.258110 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:21Z","lastTransitionTime":"2025-12-11T08:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.360469 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.360513 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.360524 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.360543 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.360553 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:21Z","lastTransitionTime":"2025-12-11T08:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.463153 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.463194 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.463207 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.463230 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.463243 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:21Z","lastTransitionTime":"2025-12-11T08:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.565785 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.565840 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.565874 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.565899 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.565914 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:21Z","lastTransitionTime":"2025-12-11T08:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.668455 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.668505 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.668517 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.668535 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.668548 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:21Z","lastTransitionTime":"2025-12-11T08:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.771185 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.771226 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.771236 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.771251 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.771260 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:21Z","lastTransitionTime":"2025-12-11T08:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.874417 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.874484 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.874500 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.874532 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.874541 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:21Z","lastTransitionTime":"2025-12-11T08:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.977127 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.977200 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.977224 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.977256 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:21 crc kubenswrapper[4629]: I1211 08:38:21.977280 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:21Z","lastTransitionTime":"2025-12-11T08:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.080230 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.080309 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.080343 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.080369 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.080381 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:22Z","lastTransitionTime":"2025-12-11T08:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.183083 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.183140 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.183153 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.183174 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.183192 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:22Z","lastTransitionTime":"2025-12-11T08:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.198279 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.198322 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.198356 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.198279 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:22 crc kubenswrapper[4629]: E1211 08:38:22.198447 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:22 crc kubenswrapper[4629]: E1211 08:38:22.198544 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:22 crc kubenswrapper[4629]: E1211 08:38:22.198619 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:22 crc kubenswrapper[4629]: E1211 08:38:22.198776 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.211704 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12aee1262818b14ec65288dd309a9b3b42fc33e7881a996fc1091449109eed3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.225477 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0693c14e-aca1-4ada-8993-b17e90f51798\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7fe01d201b10dddc135838b82a87978edca2435bcba37fcb310eea879ae143f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2590923465b1aa6abfd9ecc583cf4674e655b7f43d33aad28e0508ad2113b9c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7eab8a4edde1b300f6690beb51dc53fa4210b6ed14ae7ec40788be3f7e09995\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2561c06e9ee060777c13b0c0d6cf95bb8969dec2e8b27a8d636547dcacdae581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f30e1f40e1933444ead35f853403db77fbd08dd2b4cebc92c761587816d3113\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7febebbb123978c49d661bc949df3d677ad7ea14d4fd3ffefb5e63417587e073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://038a2e96ac0a5c0b533f4d3efecf3185cd30172488143cbc0328d20ef5944c90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqb45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gn7n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.241766 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79696728-82a3-4065-b645-cfefc699b00f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b106347064930bb61504a7d778137f8d119f631b0ce32d3204fe1dca77626956\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a4356c0b3b921157da0be896a269d3281ff9700b3932978f06838dec91bc7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jvs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-grzg4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.253414 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n44pt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e908f308-e6e4-4a2e-ab9f-48e98698f2f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e9b08695e0cdd760ecf78200940efb1971090dde33b5cf9e89ce2aab0eb11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-57vgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n44pt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.273741 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b11ff56a-d91a-492e-9707-0ddeddce3d81\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc683f85aa584b70cb018a2e744e217f1f1f1851a695f3786a8762c11ed2b8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71e1f4bdb616de4922619e280755cc10db0d21d9fe5a71904083294cab761c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0565ae9ed907520dc1de48af01d76814c46ac6c75ec4a3cac52999f75461fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6cf3a02c01344356198c2facd4879f8592dd995b07fc0018330a9f2aed607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37fa509139c4631ff77e0d4e1b1a5c3a2175eb77e3a8a55a858318d23f522e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34e17b01091f52c691e5c77d9e8d9f43deed63ea96011a76c5bc67b672a12e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f5449b30a650c00efd47406350cd1ae68bf1894bb6eb8f60153a3efb30851ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://538697c41fde757753491a92dc1ed213b71cba033cdaac2a880d040cff4b23c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.286231 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.286272 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.286281 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.286296 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.286306 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:22Z","lastTransitionTime":"2025-12-11T08:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.287692 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37701d38-a859-4d67-9257-a054c616bc26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:37:04.643432 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:37:04.644796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2167276594/tls.crt::/tmp/serving-cert-2167276594/tls.key\\\\\\\"\\\\nI1211 08:37:10.074694 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:37:10.076984 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:37:10.077037 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:37:10.077092 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:37:10.077121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:37:10.081488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 08:37:10.081520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081524 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:37:10.081528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:37:10.081531 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:37:10.081534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:37:10.081537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 08:37:10.081590 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 08:37:10.083186 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.301744 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://798c2ca122caca0359f623f24f498270743aa0fcbb553beea2466e678b49eb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.314583 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78a707d7e223066e7e39b6527af88ad22f95937c9e596bc4415333625e0327ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://425dd559e64c39941525a63b8f9eecc1ecb0f552522f8bdadec9379af88f88dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.329702 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b3e45c6-d15e-4bd9-a82b-077f00b17ace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b492a8a8919937c59d529699d56228ea963c95c80a7584319083b21145a04efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a5ed942dac8275418bcfd92da1158c628f2637b69efbaf25e357c7bd7b7de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81de2189b2524b103270ea8da0e0971c384a6e5deeecdc795cfcdfc61c32b0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.344754 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.362607 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"847b50b9-0093-404f-b0dd-63c814d8986b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfef1f9eed6e04b714a9e14a6e875e9294b9114bfbc41158ca2c3ad4f6613e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-549hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wbld9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.375271 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xq9dz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa83f85c-0050-4297-b65c-4837ecaf91b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3cdafc1c1e16f27da57d96cb7301c67dfd6230490591b50ffc3d4ad39e847c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55vjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xq9dz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.391445 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lvjmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5568b3d-2e9b-4038-bb16-4ad3d87435ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:38:00Z\\\",\\\"message\\\":\\\"2025-12-11T08:37:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6\\\\n2025-12-11T08:37:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9e7462cf-e146-42e9-be0f-8853738f74d6 to /host/opt/cni/bin/\\\\n2025-12-11T08:37:15Z [verbose] multus-daemon started\\\\n2025-12-11T08:37:15Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:38:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfwqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lvjmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.391689 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.391720 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.392829 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.392893 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.392917 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:22Z","lastTransitionTime":"2025-12-11T08:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.412189 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89bfa10-15c0-40cd-b256-31a9934c855d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:38:08Z\\\",\\\"message\\\":\\\"openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:08Z is after 2025-08-24T17:21:41Z]\\\\nI1211 08:38:08.231297 6582 kube.go:317] Updating pod openshift-multus/network-metrics-daemon-29fct\\\\nI1211 08:38:08.231275 6582 services_controller.go:434] Service openshift-dns-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-dns-operator 4bf7a6e2-037e-4e09-ad6b-2e7f1059a532 4106 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:dns-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc000701b47 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: dns-ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:38:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwkf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lt499\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.424747 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-29fct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e996ab-2163-44ad-9ecc-052a44f0da4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kbz8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:37:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-29fct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.435715 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0e625b2-b7fd-4aed-a513-8ad7f8c5bd52\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8edcf812544fff27a0105abbaae88e57c68b9823483aa578a21e79ccd342c8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://763ac579712e110e6989f12f4e7ab5ed267017bae7b82f3a28b0af7a92274bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://763ac579712e110e6989f12f4e7ab5ed267017bae7b82f3a28b0af7a92274bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.447164 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87641e0f-22fd-4027-ae0c-2250392ae1ba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://928c6f4e950677f3dfaf07954796f3afe5a1862ac10188c6a6ed7058b57c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fda4882768aefea5343885f234191e191521c54daed31d21f45121b63e09007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c7136d8cca244d250a31ffac8825125bf73e5a7bad253763a02ec9944892ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6746f274d87e9286647a6bf6a1c12cd459fe12a7ffad9891d7c1bbdef5fa4c09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:36:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:36:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:36:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.459775 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.474392 4629 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:37:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.495530 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.495601 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.495611 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.495624 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.495634 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:22Z","lastTransitionTime":"2025-12-11T08:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.597466 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.597499 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.597507 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.597520 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.597529 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:22Z","lastTransitionTime":"2025-12-11T08:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.699445 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.699487 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.699514 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.699532 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.699543 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:22Z","lastTransitionTime":"2025-12-11T08:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.802138 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.802191 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.802200 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.802217 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.802226 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:22Z","lastTransitionTime":"2025-12-11T08:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.905094 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.905127 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.905135 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.905148 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:22 crc kubenswrapper[4629]: I1211 08:38:22.905157 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:22Z","lastTransitionTime":"2025-12-11T08:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.007543 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.007582 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.007592 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.007639 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.007650 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:23Z","lastTransitionTime":"2025-12-11T08:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.110371 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.110416 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.110427 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.110442 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.110454 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:23Z","lastTransitionTime":"2025-12-11T08:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.213876 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.213925 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.213934 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.213949 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.213959 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:23Z","lastTransitionTime":"2025-12-11T08:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.316322 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.316358 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.316366 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.316382 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.316413 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:23Z","lastTransitionTime":"2025-12-11T08:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.418632 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.418673 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.418682 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.418698 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.418708 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:23Z","lastTransitionTime":"2025-12-11T08:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.520971 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.521010 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.521035 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.521054 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.521062 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:23Z","lastTransitionTime":"2025-12-11T08:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.628096 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.628151 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.628169 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.628190 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.628205 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:23Z","lastTransitionTime":"2025-12-11T08:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.730662 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.730710 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.730721 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.730734 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.730743 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:23Z","lastTransitionTime":"2025-12-11T08:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.833553 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.833597 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.833613 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.833632 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.833647 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:23Z","lastTransitionTime":"2025-12-11T08:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.936461 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.936512 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.936532 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.936554 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:23 crc kubenswrapper[4629]: I1211 08:38:23.936567 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:23Z","lastTransitionTime":"2025-12-11T08:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.042074 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.042126 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.042136 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.042152 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.042168 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:24Z","lastTransitionTime":"2025-12-11T08:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.144372 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.144415 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.144429 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.144446 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.144477 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:24Z","lastTransitionTime":"2025-12-11T08:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.198919 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.198919 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.198934 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:24 crc kubenswrapper[4629]: E1211 08:38:24.199128 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.198927 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:24 crc kubenswrapper[4629]: E1211 08:38:24.199202 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:24 crc kubenswrapper[4629]: E1211 08:38:24.199294 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:24 crc kubenswrapper[4629]: E1211 08:38:24.199392 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.246917 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.246968 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.246981 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.247001 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.247016 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:24Z","lastTransitionTime":"2025-12-11T08:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.349201 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.349232 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.349240 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.349254 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.349262 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:24Z","lastTransitionTime":"2025-12-11T08:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.451612 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.451652 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.451662 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.451685 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.451696 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:24Z","lastTransitionTime":"2025-12-11T08:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.553690 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.553739 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.553752 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.553769 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.553781 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:24Z","lastTransitionTime":"2025-12-11T08:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.656052 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.656094 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.656106 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.656129 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.656140 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:24Z","lastTransitionTime":"2025-12-11T08:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.758471 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.758508 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.758516 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.758532 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.758544 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:24Z","lastTransitionTime":"2025-12-11T08:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.860602 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.860647 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.860656 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.860670 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.860680 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:24Z","lastTransitionTime":"2025-12-11T08:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.962414 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.962454 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.962465 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.962480 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:24 crc kubenswrapper[4629]: I1211 08:38:24.962490 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:24Z","lastTransitionTime":"2025-12-11T08:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.014277 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.014315 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.014327 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.014342 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.014354 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: E1211 08:38:25.027849 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.031892 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.031919 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.031928 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.031944 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.031959 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: E1211 08:38:25.043910 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.047239 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.047268 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.047278 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.047292 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.047302 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: E1211 08:38:25.058666 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.061490 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.061524 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.061535 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.061552 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.061564 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: E1211 08:38:25.073384 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.076547 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.076580 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.076591 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.076606 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.076617 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: E1211 08:38:25.087875 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:38:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4e353408-98d5-48a9-a921-91df032ed1ab\\\",\\\"systemUUID\\\":\\\"374fbfe6-2fa2-4e29-98f2-82a5ae0d8ca0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 11 08:38:25 crc kubenswrapper[4629]: E1211 08:38:25.087997 4629 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.089209 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.089239 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.089247 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.089260 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.089269 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.192045 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.192085 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.192096 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.192111 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.192121 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.294835 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.294895 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.294907 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.294922 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.294935 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.398321 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.398494 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.398506 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.398523 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.398532 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.500382 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.500426 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.500438 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.500453 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.500464 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.603434 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.603487 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.603499 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.603517 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.603529 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.706575 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.706631 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.706644 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.706713 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.706729 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.809181 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.809227 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.809236 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.809254 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.809264 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.911337 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.911388 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.911405 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.911428 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:25 crc kubenswrapper[4629]: I1211 08:38:25.911444 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:25Z","lastTransitionTime":"2025-12-11T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.013463 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.013504 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.013515 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.013530 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.013542 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:26Z","lastTransitionTime":"2025-12-11T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.116596 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.116642 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.116652 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.116667 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.116676 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:26Z","lastTransitionTime":"2025-12-11T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.198561 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.198647 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:26 crc kubenswrapper[4629]: E1211 08:38:26.198713 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.198738 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:26 crc kubenswrapper[4629]: E1211 08:38:26.198836 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.198887 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:26 crc kubenswrapper[4629]: E1211 08:38:26.198944 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:26 crc kubenswrapper[4629]: E1211 08:38:26.199007 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.219235 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.219261 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.219271 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.219284 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.219293 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:26Z","lastTransitionTime":"2025-12-11T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.322097 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.322131 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.322147 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.322161 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.322172 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:26Z","lastTransitionTime":"2025-12-11T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.424282 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.424342 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.424361 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.424380 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.424430 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:26Z","lastTransitionTime":"2025-12-11T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.527750 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.527805 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.527817 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.527835 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.527873 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:26Z","lastTransitionTime":"2025-12-11T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.630057 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.630099 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.630108 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.630122 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.630132 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:26Z","lastTransitionTime":"2025-12-11T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.731725 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.731768 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.731776 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.731792 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.731803 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:26Z","lastTransitionTime":"2025-12-11T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.834556 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.834596 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.834607 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.834623 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.834635 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:26Z","lastTransitionTime":"2025-12-11T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.936658 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.936705 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.936714 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.936727 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:26 crc kubenswrapper[4629]: I1211 08:38:26.936736 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:26Z","lastTransitionTime":"2025-12-11T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.039424 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.039518 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.039530 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.039546 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.039557 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:27Z","lastTransitionTime":"2025-12-11T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.142154 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.142201 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.142211 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.142236 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.142247 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:27Z","lastTransitionTime":"2025-12-11T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.244960 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.244997 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.245008 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.245025 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.245036 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:27Z","lastTransitionTime":"2025-12-11T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.347338 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.347429 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.347451 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.347467 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.347479 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:27Z","lastTransitionTime":"2025-12-11T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.450368 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.450423 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.450439 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.450463 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.450480 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:27Z","lastTransitionTime":"2025-12-11T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.552707 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.552744 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.552769 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.552785 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.552795 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:27Z","lastTransitionTime":"2025-12-11T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.655286 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.655392 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.655414 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.655437 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.655452 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:27Z","lastTransitionTime":"2025-12-11T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.757901 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.757950 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.757962 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.757979 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.757993 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:27Z","lastTransitionTime":"2025-12-11T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.860442 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.860485 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.860495 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.860562 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.860572 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:27Z","lastTransitionTime":"2025-12-11T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.962697 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.962729 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.962738 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.962751 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:27 crc kubenswrapper[4629]: I1211 08:38:27.962763 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:27Z","lastTransitionTime":"2025-12-11T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.065609 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.065643 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.065653 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.065668 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.065679 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:28Z","lastTransitionTime":"2025-12-11T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.168495 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.168537 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.168546 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.168562 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.168570 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:28Z","lastTransitionTime":"2025-12-11T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.198315 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.198372 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.198405 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.198436 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:28 crc kubenswrapper[4629]: E1211 08:38:28.198476 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:28 crc kubenswrapper[4629]: E1211 08:38:28.198555 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:28 crc kubenswrapper[4629]: E1211 08:38:28.198646 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:28 crc kubenswrapper[4629]: E1211 08:38:28.198735 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.271219 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.271260 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.271270 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.271285 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.271296 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:28Z","lastTransitionTime":"2025-12-11T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.373208 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.373244 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.373253 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.373266 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.373277 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:28Z","lastTransitionTime":"2025-12-11T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.475285 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.475332 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.475349 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.475371 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.475389 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:28Z","lastTransitionTime":"2025-12-11T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.577138 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.577185 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.577202 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.577219 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.577230 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:28Z","lastTransitionTime":"2025-12-11T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.679350 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.679385 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.679394 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.679409 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.679419 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:28Z","lastTransitionTime":"2025-12-11T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.782436 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.782507 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.782519 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.782542 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.782560 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:28Z","lastTransitionTime":"2025-12-11T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.885210 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.885254 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.885265 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.885282 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.885291 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:28Z","lastTransitionTime":"2025-12-11T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.987871 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.987929 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.987941 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.987959 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:28 crc kubenswrapper[4629]: I1211 08:38:28.987970 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:28Z","lastTransitionTime":"2025-12-11T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.090147 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.090237 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.090253 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.090269 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.090281 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:29Z","lastTransitionTime":"2025-12-11T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.192764 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.192799 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.192807 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.192822 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.192831 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:29Z","lastTransitionTime":"2025-12-11T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.295533 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.295585 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.295595 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.295608 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.295617 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:29Z","lastTransitionTime":"2025-12-11T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.398303 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.398344 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.398352 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.398367 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.398376 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:29Z","lastTransitionTime":"2025-12-11T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.500290 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.500335 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.500348 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.500366 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.500379 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:29Z","lastTransitionTime":"2025-12-11T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.602005 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.602039 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.602048 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.602062 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.602071 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:29Z","lastTransitionTime":"2025-12-11T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.704341 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.704397 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.704411 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.704429 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.704478 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:29Z","lastTransitionTime":"2025-12-11T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.806734 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.806790 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.806802 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.806817 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.806830 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:29Z","lastTransitionTime":"2025-12-11T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.909084 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.909150 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.909169 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.909187 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:29 crc kubenswrapper[4629]: I1211 08:38:29.909205 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:29Z","lastTransitionTime":"2025-12-11T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.012318 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.012379 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.012391 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.012411 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.012424 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:30Z","lastTransitionTime":"2025-12-11T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.106712 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:30 crc kubenswrapper[4629]: E1211 08:38:30.107014 4629 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:38:30 crc kubenswrapper[4629]: E1211 08:38:30.107183 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs podName:13e996ab-2163-44ad-9ecc-052a44f0da4c nodeName:}" failed. No retries permitted until 2025-12-11 08:39:34.107156352 +0000 UTC m=+162.172572060 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs") pod "network-metrics-daemon-29fct" (UID: "13e996ab-2163-44ad-9ecc-052a44f0da4c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.116041 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.116093 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.116110 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.116144 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.116162 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:30Z","lastTransitionTime":"2025-12-11T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.199005 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.199114 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:30 crc kubenswrapper[4629]: E1211 08:38:30.199181 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.199199 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:30 crc kubenswrapper[4629]: E1211 08:38:30.199339 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.199356 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:30 crc kubenswrapper[4629]: E1211 08:38:30.199603 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:30 crc kubenswrapper[4629]: E1211 08:38:30.199633 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.219120 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.219148 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.219159 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.219173 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.219184 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:30Z","lastTransitionTime":"2025-12-11T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.321953 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.321990 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.322002 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.322025 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.322037 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:30Z","lastTransitionTime":"2025-12-11T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.424533 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.424570 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.424587 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.424606 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.424617 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:30Z","lastTransitionTime":"2025-12-11T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.530544 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.530583 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.530591 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.530605 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.530614 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:30Z","lastTransitionTime":"2025-12-11T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.633460 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.633518 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.633527 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.633539 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.633565 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:30Z","lastTransitionTime":"2025-12-11T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.736122 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.736175 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.736188 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.736402 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.736416 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:30Z","lastTransitionTime":"2025-12-11T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.838623 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.838653 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.838662 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.838675 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.838701 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:30Z","lastTransitionTime":"2025-12-11T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.949746 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.949784 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.949794 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.949810 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:30 crc kubenswrapper[4629]: I1211 08:38:30.949822 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:30Z","lastTransitionTime":"2025-12-11T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.053239 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.053285 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.053299 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.053317 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.053331 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:31Z","lastTransitionTime":"2025-12-11T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.156373 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.156424 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.156441 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.156460 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.156471 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:31Z","lastTransitionTime":"2025-12-11T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.259294 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.259328 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.259336 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.259348 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.259358 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:31Z","lastTransitionTime":"2025-12-11T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.361348 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.361387 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.361397 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.361411 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.361420 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:31Z","lastTransitionTime":"2025-12-11T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.463986 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.464027 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.464039 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.464055 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.464066 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:31Z","lastTransitionTime":"2025-12-11T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.566370 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.566414 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.566422 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.566437 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.566447 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:31Z","lastTransitionTime":"2025-12-11T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.669278 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.669325 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.669353 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.669378 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.669393 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:31Z","lastTransitionTime":"2025-12-11T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.772364 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.772407 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.772417 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.772432 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.772464 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:31Z","lastTransitionTime":"2025-12-11T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.875153 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.875218 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.875235 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.875262 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.875299 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:31Z","lastTransitionTime":"2025-12-11T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.977687 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.977744 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.977762 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.977784 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:31 crc kubenswrapper[4629]: I1211 08:38:31.977800 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:31Z","lastTransitionTime":"2025-12-11T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.080549 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.080600 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.080614 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.080636 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.080653 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:32Z","lastTransitionTime":"2025-12-11T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.183327 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.183368 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.183380 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.183395 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.183405 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:32Z","lastTransitionTime":"2025-12-11T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.198821 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.198932 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.198835 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:32 crc kubenswrapper[4629]: E1211 08:38:32.199015 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.199045 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:32 crc kubenswrapper[4629]: E1211 08:38:32.199117 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:32 crc kubenswrapper[4629]: E1211 08:38:32.199175 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:32 crc kubenswrapper[4629]: E1211 08:38:32.199242 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.199942 4629 scope.go:117] "RemoveContainer" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" Dec 11 08:38:32 crc kubenswrapper[4629]: E1211 08:38:32.200340 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.215750 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=76.215712898 podStartE2EDuration="1m16.215712898s" podCreationTimestamp="2025-12-11 08:37:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:32.215317226 +0000 UTC m=+100.280732844" watchObservedRunningTime="2025-12-11 08:38:32.215712898 +0000 UTC m=+100.281128516" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.262707 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-lvjmx" podStartSLOduration=81.262682819 podStartE2EDuration="1m21.262682819s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:32.261927896 +0000 UTC m=+100.327343524" watchObservedRunningTime="2025-12-11 08:38:32.262682819 +0000 UTC m=+100.328098447" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.263040 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podStartSLOduration=81.26302961 podStartE2EDuration="1m21.26302961s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:32.245307315 +0000 UTC m=+100.310722953" watchObservedRunningTime="2025-12-11 08:38:32.26302961 +0000 UTC m=+100.328445228" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.286696 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.286745 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.286755 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.286772 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.286790 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:32Z","lastTransitionTime":"2025-12-11T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.315371 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=22.315350473 podStartE2EDuration="22.315350473s" podCreationTimestamp="2025-12-11 08:38:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:32.314790295 +0000 UTC m=+100.380205903" watchObservedRunningTime="2025-12-11 08:38:32.315350473 +0000 UTC m=+100.380766081" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.331690 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=53.331670264 podStartE2EDuration="53.331670264s" podCreationTimestamp="2025-12-11 08:37:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:32.330750485 +0000 UTC m=+100.396166103" watchObservedRunningTime="2025-12-11 08:38:32.331670264 +0000 UTC m=+100.397085872" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.381345 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-xq9dz" podStartSLOduration=81.381327121 podStartE2EDuration="1m21.381327121s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:32.368872173 +0000 UTC m=+100.434287781" watchObservedRunningTime="2025-12-11 08:38:32.381327121 +0000 UTC m=+100.446742719" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.388895 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.388924 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.388947 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.388963 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.388973 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:32Z","lastTransitionTime":"2025-12-11T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.399052 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-gn7n4" podStartSLOduration=81.399037927 podStartE2EDuration="1m21.399037927s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:32.398891382 +0000 UTC m=+100.464306990" watchObservedRunningTime="2025-12-11 08:38:32.399037927 +0000 UTC m=+100.464453535" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.441365 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=80.441329438 podStartE2EDuration="1m20.441329438s" podCreationTimestamp="2025-12-11 08:37:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:32.440602625 +0000 UTC m=+100.506018253" watchObservedRunningTime="2025-12-11 08:38:32.441329438 +0000 UTC m=+100.506745056" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.441674 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-grzg4" podStartSLOduration=81.441651519 podStartE2EDuration="1m21.441651519s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:32.412979462 +0000 UTC m=+100.478395070" watchObservedRunningTime="2025-12-11 08:38:32.441651519 +0000 UTC m=+100.507067127" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.460392 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=82.460371217 podStartE2EDuration="1m22.460371217s" podCreationTimestamp="2025-12-11 08:37:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:32.457676231 +0000 UTC m=+100.523091839" watchObservedRunningTime="2025-12-11 08:38:32.460371217 +0000 UTC m=+100.525786835" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.491260 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.491315 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.491325 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.491341 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.491375 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:32Z","lastTransitionTime":"2025-12-11T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.594281 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.594358 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.594368 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.594385 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.594395 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:32Z","lastTransitionTime":"2025-12-11T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.696718 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.696764 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.696776 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.696791 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.696800 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:32Z","lastTransitionTime":"2025-12-11T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.798528 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.798566 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.798578 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.798594 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.798604 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:32Z","lastTransitionTime":"2025-12-11T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.900989 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.901030 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.901040 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.901058 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:32 crc kubenswrapper[4629]: I1211 08:38:32.901067 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:32Z","lastTransitionTime":"2025-12-11T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.003187 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.003232 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.003243 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.003270 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.003282 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:33Z","lastTransitionTime":"2025-12-11T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.105730 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.105791 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.105809 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.105835 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.105883 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:33Z","lastTransitionTime":"2025-12-11T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.208143 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.208178 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.208194 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.208210 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.208224 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:33Z","lastTransitionTime":"2025-12-11T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.313243 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.313282 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.313292 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.313305 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.313314 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:33Z","lastTransitionTime":"2025-12-11T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.415305 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.415346 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.415358 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.415375 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.415389 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:33Z","lastTransitionTime":"2025-12-11T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.517521 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.517566 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.517576 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.517591 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.517600 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:33Z","lastTransitionTime":"2025-12-11T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.619946 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.619994 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.620005 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.620023 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.620035 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:33Z","lastTransitionTime":"2025-12-11T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.722769 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.722817 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.722829 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.722883 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.722902 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:33Z","lastTransitionTime":"2025-12-11T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.825129 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.825181 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.825206 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.825226 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.825241 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:33Z","lastTransitionTime":"2025-12-11T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.927216 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.927256 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.927264 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.927279 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:33 crc kubenswrapper[4629]: I1211 08:38:33.927288 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:33Z","lastTransitionTime":"2025-12-11T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.030009 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.030081 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.030099 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.030117 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.030129 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:34Z","lastTransitionTime":"2025-12-11T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.132421 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.132470 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.132480 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.132498 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.132511 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:34Z","lastTransitionTime":"2025-12-11T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.198459 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.198514 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.198526 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:34 crc kubenswrapper[4629]: E1211 08:38:34.198603 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.198612 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:34 crc kubenswrapper[4629]: E1211 08:38:34.198683 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:34 crc kubenswrapper[4629]: E1211 08:38:34.198815 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:34 crc kubenswrapper[4629]: E1211 08:38:34.198915 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.234550 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.234588 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.234615 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.234629 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.234638 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:34Z","lastTransitionTime":"2025-12-11T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.336895 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.336923 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.336931 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.336942 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.336952 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:34Z","lastTransitionTime":"2025-12-11T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.439591 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.439630 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.439640 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.439656 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.439667 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:34Z","lastTransitionTime":"2025-12-11T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.542638 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.542691 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.542705 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.542723 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.542736 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:34Z","lastTransitionTime":"2025-12-11T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.645993 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.646067 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.646081 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.646103 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.646148 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:34Z","lastTransitionTime":"2025-12-11T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.749502 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.749566 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.749586 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.749610 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.749627 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:34Z","lastTransitionTime":"2025-12-11T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.853587 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.853641 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.853658 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.853679 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.853694 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:34Z","lastTransitionTime":"2025-12-11T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.956382 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.956420 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.956428 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.956445 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:34 crc kubenswrapper[4629]: I1211 08:38:34.956454 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:34Z","lastTransitionTime":"2025-12-11T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.060386 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.060442 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.060458 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.060484 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.060502 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:35Z","lastTransitionTime":"2025-12-11T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.163052 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.163115 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.163131 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.163154 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.163168 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:35Z","lastTransitionTime":"2025-12-11T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.265527 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.265588 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.265596 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.265614 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.265627 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:35Z","lastTransitionTime":"2025-12-11T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.343272 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.343321 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.343334 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.343351 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.343363 4629 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:38:35Z","lastTransitionTime":"2025-12-11T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.396680 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-n44pt" podStartSLOduration=84.396662382 podStartE2EDuration="1m24.396662382s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:32.514056472 +0000 UTC m=+100.579472100" watchObservedRunningTime="2025-12-11 08:38:35.396662382 +0000 UTC m=+103.462078000" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.397402 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27"] Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.397930 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.400756 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.400797 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.401181 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.401235 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.464808 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/19197a16-8589-4e59-b58b-34e241b9e81c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.465239 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/19197a16-8589-4e59-b58b-34e241b9e81c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.465261 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19197a16-8589-4e59-b58b-34e241b9e81c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.465294 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/19197a16-8589-4e59-b58b-34e241b9e81c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.465403 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/19197a16-8589-4e59-b58b-34e241b9e81c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.566544 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/19197a16-8589-4e59-b58b-34e241b9e81c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.566638 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19197a16-8589-4e59-b58b-34e241b9e81c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.566692 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/19197a16-8589-4e59-b58b-34e241b9e81c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.566739 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/19197a16-8589-4e59-b58b-34e241b9e81c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.566784 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/19197a16-8589-4e59-b58b-34e241b9e81c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.567051 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/19197a16-8589-4e59-b58b-34e241b9e81c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.567601 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/19197a16-8589-4e59-b58b-34e241b9e81c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.570495 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/19197a16-8589-4e59-b58b-34e241b9e81c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.576827 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19197a16-8589-4e59-b58b-34e241b9e81c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.586951 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/19197a16-8589-4e59-b58b-34e241b9e81c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w9j27\" (UID: \"19197a16-8589-4e59-b58b-34e241b9e81c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:35 crc kubenswrapper[4629]: I1211 08:38:35.728165 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" Dec 11 08:38:36 crc kubenswrapper[4629]: I1211 08:38:36.198910 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:36 crc kubenswrapper[4629]: I1211 08:38:36.198914 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:36 crc kubenswrapper[4629]: I1211 08:38:36.198945 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:36 crc kubenswrapper[4629]: I1211 08:38:36.198959 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:36 crc kubenswrapper[4629]: E1211 08:38:36.199332 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:36 crc kubenswrapper[4629]: E1211 08:38:36.199727 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:36 crc kubenswrapper[4629]: E1211 08:38:36.199867 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:36 crc kubenswrapper[4629]: E1211 08:38:36.199938 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:36 crc kubenswrapper[4629]: I1211 08:38:36.759373 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" event={"ID":"19197a16-8589-4e59-b58b-34e241b9e81c","Type":"ContainerStarted","Data":"e966decb93f04678aa04dde1240e47e7d937fa72e65fdbb0378ce8c06dae5b90"} Dec 11 08:38:36 crc kubenswrapper[4629]: I1211 08:38:36.760002 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" event={"ID":"19197a16-8589-4e59-b58b-34e241b9e81c","Type":"ContainerStarted","Data":"0a922e3f9f0505c91c4b34b20dea1964476bd533b6b5179957e01c7112a5fd16"} Dec 11 08:38:36 crc kubenswrapper[4629]: I1211 08:38:36.775995 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w9j27" podStartSLOduration=85.775975937 podStartE2EDuration="1m25.775975937s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:36.775541004 +0000 UTC m=+104.840956622" watchObservedRunningTime="2025-12-11 08:38:36.775975937 +0000 UTC m=+104.841391545" Dec 11 08:38:38 crc kubenswrapper[4629]: I1211 08:38:38.198196 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:38 crc kubenswrapper[4629]: I1211 08:38:38.198271 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:38 crc kubenswrapper[4629]: I1211 08:38:38.198332 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:38 crc kubenswrapper[4629]: E1211 08:38:38.198447 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:38 crc kubenswrapper[4629]: I1211 08:38:38.198479 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:38 crc kubenswrapper[4629]: E1211 08:38:38.198626 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:38 crc kubenswrapper[4629]: E1211 08:38:38.198647 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:38 crc kubenswrapper[4629]: E1211 08:38:38.198765 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:40 crc kubenswrapper[4629]: I1211 08:38:40.198243 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:40 crc kubenswrapper[4629]: I1211 08:38:40.198282 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:40 crc kubenswrapper[4629]: I1211 08:38:40.198316 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:40 crc kubenswrapper[4629]: I1211 08:38:40.198267 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:40 crc kubenswrapper[4629]: E1211 08:38:40.198375 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:40 crc kubenswrapper[4629]: E1211 08:38:40.198439 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:40 crc kubenswrapper[4629]: E1211 08:38:40.198488 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:40 crc kubenswrapper[4629]: E1211 08:38:40.198528 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:42 crc kubenswrapper[4629]: I1211 08:38:42.198583 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:42 crc kubenswrapper[4629]: I1211 08:38:42.198596 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:42 crc kubenswrapper[4629]: I1211 08:38:42.198633 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:42 crc kubenswrapper[4629]: I1211 08:38:42.198651 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:42 crc kubenswrapper[4629]: E1211 08:38:42.199867 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:42 crc kubenswrapper[4629]: E1211 08:38:42.199920 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:42 crc kubenswrapper[4629]: E1211 08:38:42.200065 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:42 crc kubenswrapper[4629]: E1211 08:38:42.200130 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:43 crc kubenswrapper[4629]: I1211 08:38:43.198839 4629 scope.go:117] "RemoveContainer" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" Dec 11 08:38:43 crc kubenswrapper[4629]: E1211 08:38:43.199016 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lt499_openshift-ovn-kubernetes(f89bfa10-15c0-40cd-b256-31a9934c855d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" Dec 11 08:38:44 crc kubenswrapper[4629]: I1211 08:38:44.198304 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:44 crc kubenswrapper[4629]: E1211 08:38:44.198827 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:44 crc kubenswrapper[4629]: I1211 08:38:44.198370 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:44 crc kubenswrapper[4629]: E1211 08:38:44.198920 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:44 crc kubenswrapper[4629]: I1211 08:38:44.198404 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:44 crc kubenswrapper[4629]: I1211 08:38:44.198317 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:44 crc kubenswrapper[4629]: E1211 08:38:44.199171 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:44 crc kubenswrapper[4629]: E1211 08:38:44.198977 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:46 crc kubenswrapper[4629]: I1211 08:38:46.197962 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:46 crc kubenswrapper[4629]: I1211 08:38:46.198015 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:46 crc kubenswrapper[4629]: I1211 08:38:46.198021 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:46 crc kubenswrapper[4629]: I1211 08:38:46.197983 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:46 crc kubenswrapper[4629]: E1211 08:38:46.198103 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:46 crc kubenswrapper[4629]: E1211 08:38:46.198186 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:46 crc kubenswrapper[4629]: E1211 08:38:46.198253 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:46 crc kubenswrapper[4629]: E1211 08:38:46.198309 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:46 crc kubenswrapper[4629]: I1211 08:38:46.788363 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lvjmx_f5568b3d-2e9b-4038-bb16-4ad3d87435ee/kube-multus/1.log" Dec 11 08:38:46 crc kubenswrapper[4629]: I1211 08:38:46.789145 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lvjmx_f5568b3d-2e9b-4038-bb16-4ad3d87435ee/kube-multus/0.log" Dec 11 08:38:46 crc kubenswrapper[4629]: I1211 08:38:46.789302 4629 generic.go:334] "Generic (PLEG): container finished" podID="f5568b3d-2e9b-4038-bb16-4ad3d87435ee" containerID="9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6" exitCode=1 Dec 11 08:38:46 crc kubenswrapper[4629]: I1211 08:38:46.789412 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lvjmx" event={"ID":"f5568b3d-2e9b-4038-bb16-4ad3d87435ee","Type":"ContainerDied","Data":"9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6"} Dec 11 08:38:46 crc kubenswrapper[4629]: I1211 08:38:46.789648 4629 scope.go:117] "RemoveContainer" containerID="26cad97d580b3187e33c8382616ab45192d79fef34054f4dfd0aea2fb723c2ba" Dec 11 08:38:46 crc kubenswrapper[4629]: I1211 08:38:46.790413 4629 scope.go:117] "RemoveContainer" containerID="9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6" Dec 11 08:38:46 crc kubenswrapper[4629]: E1211 08:38:46.790816 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-lvjmx_openshift-multus(f5568b3d-2e9b-4038-bb16-4ad3d87435ee)\"" pod="openshift-multus/multus-lvjmx" podUID="f5568b3d-2e9b-4038-bb16-4ad3d87435ee" Dec 11 08:38:47 crc kubenswrapper[4629]: I1211 08:38:47.793397 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lvjmx_f5568b3d-2e9b-4038-bb16-4ad3d87435ee/kube-multus/1.log" Dec 11 08:38:48 crc kubenswrapper[4629]: I1211 08:38:48.198546 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:48 crc kubenswrapper[4629]: I1211 08:38:48.198554 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:48 crc kubenswrapper[4629]: I1211 08:38:48.198688 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:48 crc kubenswrapper[4629]: I1211 08:38:48.198737 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:48 crc kubenswrapper[4629]: E1211 08:38:48.198834 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:48 crc kubenswrapper[4629]: E1211 08:38:48.198963 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:48 crc kubenswrapper[4629]: E1211 08:38:48.199024 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:48 crc kubenswrapper[4629]: E1211 08:38:48.199090 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:50 crc kubenswrapper[4629]: I1211 08:38:50.198050 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:50 crc kubenswrapper[4629]: I1211 08:38:50.198085 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:50 crc kubenswrapper[4629]: I1211 08:38:50.198100 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:50 crc kubenswrapper[4629]: E1211 08:38:50.198201 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:50 crc kubenswrapper[4629]: E1211 08:38:50.198325 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:50 crc kubenswrapper[4629]: E1211 08:38:50.198423 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:50 crc kubenswrapper[4629]: I1211 08:38:50.198944 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:50 crc kubenswrapper[4629]: E1211 08:38:50.199021 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:52 crc kubenswrapper[4629]: I1211 08:38:52.197978 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:52 crc kubenswrapper[4629]: E1211 08:38:52.198118 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:52 crc kubenswrapper[4629]: I1211 08:38:52.199284 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:52 crc kubenswrapper[4629]: E1211 08:38:52.199332 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:52 crc kubenswrapper[4629]: I1211 08:38:52.199369 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:52 crc kubenswrapper[4629]: E1211 08:38:52.199423 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:52 crc kubenswrapper[4629]: I1211 08:38:52.199589 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:52 crc kubenswrapper[4629]: E1211 08:38:52.199641 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:52 crc kubenswrapper[4629]: E1211 08:38:52.218838 4629 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 11 08:38:52 crc kubenswrapper[4629]: E1211 08:38:52.324495 4629 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 11 08:38:54 crc kubenswrapper[4629]: I1211 08:38:54.198092 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:54 crc kubenswrapper[4629]: I1211 08:38:54.198159 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:54 crc kubenswrapper[4629]: I1211 08:38:54.198198 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:54 crc kubenswrapper[4629]: I1211 08:38:54.198274 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:54 crc kubenswrapper[4629]: E1211 08:38:54.198274 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:54 crc kubenswrapper[4629]: E1211 08:38:54.198530 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:54 crc kubenswrapper[4629]: E1211 08:38:54.198614 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:54 crc kubenswrapper[4629]: E1211 08:38:54.198443 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:56 crc kubenswrapper[4629]: I1211 08:38:56.198876 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:56 crc kubenswrapper[4629]: I1211 08:38:56.199237 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:56 crc kubenswrapper[4629]: I1211 08:38:56.199277 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:56 crc kubenswrapper[4629]: E1211 08:38:56.199309 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:56 crc kubenswrapper[4629]: E1211 08:38:56.199398 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:56 crc kubenswrapper[4629]: I1211 08:38:56.199447 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:56 crc kubenswrapper[4629]: E1211 08:38:56.199578 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:56 crc kubenswrapper[4629]: E1211 08:38:56.199647 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:56 crc kubenswrapper[4629]: I1211 08:38:56.200507 4629 scope.go:117] "RemoveContainer" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" Dec 11 08:38:56 crc kubenswrapper[4629]: I1211 08:38:56.823686 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/3.log" Dec 11 08:38:56 crc kubenswrapper[4629]: I1211 08:38:56.826579 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerStarted","Data":"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f"} Dec 11 08:38:56 crc kubenswrapper[4629]: I1211 08:38:56.826985 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:38:56 crc kubenswrapper[4629]: I1211 08:38:56.853646 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podStartSLOduration=105.853606129 podStartE2EDuration="1m45.853606129s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:38:56.852010108 +0000 UTC m=+124.917425716" watchObservedRunningTime="2025-12-11 08:38:56.853606129 +0000 UTC m=+124.919021737" Dec 11 08:38:57 crc kubenswrapper[4629]: I1211 08:38:57.020765 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-29fct"] Dec 11 08:38:57 crc kubenswrapper[4629]: I1211 08:38:57.020904 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:57 crc kubenswrapper[4629]: E1211 08:38:57.020992 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:38:57 crc kubenswrapper[4629]: E1211 08:38:57.325701 4629 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 11 08:38:58 crc kubenswrapper[4629]: I1211 08:38:58.198395 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:38:58 crc kubenswrapper[4629]: I1211 08:38:58.198435 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:38:58 crc kubenswrapper[4629]: E1211 08:38:58.198547 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:38:58 crc kubenswrapper[4629]: I1211 08:38:58.198568 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:38:58 crc kubenswrapper[4629]: E1211 08:38:58.198616 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:38:58 crc kubenswrapper[4629]: E1211 08:38:58.199138 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:38:59 crc kubenswrapper[4629]: I1211 08:38:59.198295 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:38:59 crc kubenswrapper[4629]: E1211 08:38:59.198451 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:39:00 crc kubenswrapper[4629]: I1211 08:39:00.198590 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:39:00 crc kubenswrapper[4629]: I1211 08:39:00.198647 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:39:00 crc kubenswrapper[4629]: E1211 08:39:00.198772 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:39:00 crc kubenswrapper[4629]: E1211 08:39:00.198903 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:39:00 crc kubenswrapper[4629]: I1211 08:39:00.199099 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:39:00 crc kubenswrapper[4629]: E1211 08:39:00.199179 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:39:01 crc kubenswrapper[4629]: I1211 08:39:01.198630 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:39:01 crc kubenswrapper[4629]: E1211 08:39:01.199062 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:39:01 crc kubenswrapper[4629]: I1211 08:39:01.199193 4629 scope.go:117] "RemoveContainer" containerID="9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6" Dec 11 08:39:01 crc kubenswrapper[4629]: I1211 08:39:01.843054 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lvjmx_f5568b3d-2e9b-4038-bb16-4ad3d87435ee/kube-multus/1.log" Dec 11 08:39:01 crc kubenswrapper[4629]: I1211 08:39:01.843444 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lvjmx" event={"ID":"f5568b3d-2e9b-4038-bb16-4ad3d87435ee","Type":"ContainerStarted","Data":"a1c5e5255a2e972dcbb5f2f48978a4d9c479227775a20539ef2106ea7eb29d01"} Dec 11 08:39:02 crc kubenswrapper[4629]: I1211 08:39:02.198039 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:39:02 crc kubenswrapper[4629]: I1211 08:39:02.198086 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:39:02 crc kubenswrapper[4629]: E1211 08:39:02.199628 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:39:02 crc kubenswrapper[4629]: I1211 08:39:02.199708 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:39:02 crc kubenswrapper[4629]: E1211 08:39:02.199785 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:39:02 crc kubenswrapper[4629]: E1211 08:39:02.199936 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:39:02 crc kubenswrapper[4629]: E1211 08:39:02.326452 4629 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 11 08:39:03 crc kubenswrapper[4629]: I1211 08:39:03.198914 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:39:03 crc kubenswrapper[4629]: E1211 08:39:03.199062 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:39:04 crc kubenswrapper[4629]: I1211 08:39:04.198753 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:39:04 crc kubenswrapper[4629]: I1211 08:39:04.198788 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:39:04 crc kubenswrapper[4629]: I1211 08:39:04.198798 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:39:04 crc kubenswrapper[4629]: E1211 08:39:04.198979 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:39:04 crc kubenswrapper[4629]: E1211 08:39:04.199194 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:39:04 crc kubenswrapper[4629]: E1211 08:39:04.199334 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:39:05 crc kubenswrapper[4629]: I1211 08:39:05.198828 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:39:05 crc kubenswrapper[4629]: E1211 08:39:05.198988 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:39:06 crc kubenswrapper[4629]: I1211 08:39:06.198763 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:39:06 crc kubenswrapper[4629]: I1211 08:39:06.198816 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:39:06 crc kubenswrapper[4629]: I1211 08:39:06.198880 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:39:06 crc kubenswrapper[4629]: E1211 08:39:06.199918 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:39:06 crc kubenswrapper[4629]: E1211 08:39:06.199990 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:39:06 crc kubenswrapper[4629]: E1211 08:39:06.200007 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:39:07 crc kubenswrapper[4629]: I1211 08:39:07.197948 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:39:07 crc kubenswrapper[4629]: E1211 08:39:07.198085 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-29fct" podUID="13e996ab-2163-44ad-9ecc-052a44f0da4c" Dec 11 08:39:08 crc kubenswrapper[4629]: I1211 08:39:08.198249 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:39:08 crc kubenswrapper[4629]: I1211 08:39:08.198261 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:39:08 crc kubenswrapper[4629]: I1211 08:39:08.198310 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:39:08 crc kubenswrapper[4629]: I1211 08:39:08.203518 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 11 08:39:08 crc kubenswrapper[4629]: I1211 08:39:08.204687 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 11 08:39:08 crc kubenswrapper[4629]: I1211 08:39:08.204787 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 11 08:39:08 crc kubenswrapper[4629]: I1211 08:39:08.205193 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 11 08:39:09 crc kubenswrapper[4629]: I1211 08:39:09.198042 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:39:09 crc kubenswrapper[4629]: I1211 08:39:09.201051 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 11 08:39:09 crc kubenswrapper[4629]: I1211 08:39:09.201145 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 11 08:39:10 crc kubenswrapper[4629]: I1211 08:39:10.772548 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.464259 4629 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.506646 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t7tg7"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.507371 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.507892 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-npk4p"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.516445 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.517354 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.532949 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bktzv"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.533252 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.533297 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.533533 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.533594 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.533650 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.533748 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.534059 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.534567 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.535663 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.536793 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2c7tp"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.537301 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.537903 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.538414 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.545950 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-lhszp"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.546591 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.551396 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.552222 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.552283 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cmklq"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.553583 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.559500 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.559776 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.559933 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.560264 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.560435 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.560616 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.560766 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.566609 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.566686 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.566943 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.566936 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.567133 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.567389 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.567563 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.568204 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.569208 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.569836 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.569253 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.575220 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4jtwf"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.575897 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.577428 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.577593 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.577771 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.577958 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.577981 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.578101 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.578270 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.578310 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.578414 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.578815 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.578981 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.579389 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.579499 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.579631 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.579725 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.579818 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.579988 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.580091 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.580182 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.580206 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.580426 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.580100 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.581004 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.581139 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.578273 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.581244 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.581352 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.581432 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.581739 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.581794 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.581941 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.582179 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.582343 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.582505 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.583582 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.584122 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.584751 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.585017 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.587240 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.589561 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.607939 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.608231 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8q9kf"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.608438 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.608622 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.608872 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.608900 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.609158 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.610579 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.611045 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.611581 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.611918 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.639633 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-wbrg7"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.640279 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.640348 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.640485 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.640545 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.640651 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.640490 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.640648 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.640876 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.641209 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc12104e-ef88-4e7f-aae8-89fdd6641c7e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-t7dc6\" (UID: \"bc12104e-ef88-4e7f-aae8-89fdd6641c7e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.641298 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb8zq\" (UniqueName: \"kubernetes.io/projected/bc12104e-ef88-4e7f-aae8-89fdd6641c7e-kube-api-access-jb8zq\") pod \"cluster-image-registry-operator-dc59b4c8b-t7dc6\" (UID: \"bc12104e-ef88-4e7f-aae8-89fdd6641c7e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.641327 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc12104e-ef88-4e7f-aae8-89fdd6641c7e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-t7dc6\" (UID: \"bc12104e-ef88-4e7f-aae8-89fdd6641c7e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.641373 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc12104e-ef88-4e7f-aae8-89fdd6641c7e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-t7dc6\" (UID: \"bc12104e-ef88-4e7f-aae8-89fdd6641c7e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.641661 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.642806 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.644305 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.645432 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.646054 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.646963 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.647268 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.647390 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.647038 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.647813 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-npk4p"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.648062 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.648954 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bktzv"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.656302 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.649065 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-wbrg7" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.648485 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.648592 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.649240 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.649311 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.649382 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.649481 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.658285 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.661747 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t7tg7"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.662419 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5ccrx"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.663402 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-5ccrx" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.664099 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.664763 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.666423 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-lhszp"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.666865 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.669219 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.669292 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.669458 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.669356 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.669480 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.669972 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.670096 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.670221 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.670275 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.670368 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.670751 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.671215 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.671336 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.673335 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.678979 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-wbrg7"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.679067 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.679997 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4jtwf"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.709412 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.709991 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.710554 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.713208 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2c7tp"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.729921 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-r9x7m"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.732116 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.735802 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.736027 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.737034 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.741572 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nn9r8"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.743012 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.744007 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6h74\" (UniqueName: \"kubernetes.io/projected/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-kube-api-access-b6h74\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.744046 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-config\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.744094 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb8zq\" (UniqueName: \"kubernetes.io/projected/bc12104e-ef88-4e7f-aae8-89fdd6641c7e-kube-api-access-jb8zq\") pod \"cluster-image-registry-operator-dc59b4c8b-t7dc6\" (UID: \"bc12104e-ef88-4e7f-aae8-89fdd6641c7e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.744122 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9167ddcb-69c2-4861-80e2-1ab4419b8e93-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bnswd\" (UID: \"9167ddcb-69c2-4861-80e2-1ab4419b8e93\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.744143 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvp97\" (UniqueName: \"kubernetes.io/projected/9167ddcb-69c2-4861-80e2-1ab4419b8e93-kube-api-access-zvp97\") pod \"openshift-apiserver-operator-796bbdcf4f-bnswd\" (UID: \"9167ddcb-69c2-4861-80e2-1ab4419b8e93\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.744186 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.744206 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9167ddcb-69c2-4861-80e2-1ab4419b8e93-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bnswd\" (UID: \"9167ddcb-69c2-4861-80e2-1ab4419b8e93\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.744252 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-service-ca-bundle\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.744296 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc12104e-ef88-4e7f-aae8-89fdd6641c7e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-t7dc6\" (UID: \"bc12104e-ef88-4e7f-aae8-89fdd6641c7e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.744324 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc12104e-ef88-4e7f-aae8-89fdd6641c7e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-t7dc6\" (UID: \"bc12104e-ef88-4e7f-aae8-89fdd6641c7e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.744348 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc12104e-ef88-4e7f-aae8-89fdd6641c7e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-t7dc6\" (UID: \"bc12104e-ef88-4e7f-aae8-89fdd6641c7e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.744374 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-serving-cert\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.746946 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc12104e-ef88-4e7f-aae8-89fdd6641c7e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-t7dc6\" (UID: \"bc12104e-ef88-4e7f-aae8-89fdd6641c7e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.747566 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.748614 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.751630 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.753692 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.783641 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.783910 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-mbkcz"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.784428 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.784884 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.785122 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mbkcz" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.792598 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc12104e-ef88-4e7f-aae8-89fdd6641c7e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-t7dc6\" (UID: \"bc12104e-ef88-4e7f-aae8-89fdd6641c7e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.795161 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5ccrx"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.795250 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8q9kf"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.795262 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.795278 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.795723 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.796121 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.796458 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.796800 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.797078 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.797291 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.797382 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.797447 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.802480 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.803122 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.804030 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.809930 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.809984 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.810392 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fmmgm"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.812482 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.812674 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.813506 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nn9r8"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.813606 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.813616 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.814206 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.816755 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.817218 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rzqjs"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.817499 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.817823 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.818291 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.818430 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.818944 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fmmgm" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.828310 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.828363 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6qgqv"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.829007 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.833682 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.834149 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.834487 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-rzxln"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.834940 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rzxln" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.835210 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.839926 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.840268 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.844069 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.844121 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.844132 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cmklq"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.845044 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6h74\" (UniqueName: \"kubernetes.io/projected/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-kube-api-access-b6h74\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.845085 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-config\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.845195 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9167ddcb-69c2-4861-80e2-1ab4419b8e93-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bnswd\" (UID: \"9167ddcb-69c2-4861-80e2-1ab4419b8e93\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.845214 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvp97\" (UniqueName: \"kubernetes.io/projected/9167ddcb-69c2-4861-80e2-1ab4419b8e93-kube-api-access-zvp97\") pod \"openshift-apiserver-operator-796bbdcf4f-bnswd\" (UID: \"9167ddcb-69c2-4861-80e2-1ab4419b8e93\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.845234 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.845250 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9167ddcb-69c2-4861-80e2-1ab4419b8e93-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bnswd\" (UID: \"9167ddcb-69c2-4861-80e2-1ab4419b8e93\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.845267 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-service-ca-bundle\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.845316 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-serving-cert\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.847341 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-config\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.848677 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-service-ca-bundle\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.848825 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9167ddcb-69c2-4861-80e2-1ab4419b8e93-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bnswd\" (UID: \"9167ddcb-69c2-4861-80e2-1ab4419b8e93\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.849605 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.855381 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9167ddcb-69c2-4861-80e2-1ab4419b8e93-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bnswd\" (UID: \"9167ddcb-69c2-4861-80e2-1ab4419b8e93\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.855438 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.856698 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.858038 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fmmgm"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.861476 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-n5p54"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.873070 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rzqjs"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.873116 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.873149 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rzxln"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.873163 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wdcs7"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.873421 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.874405 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-mbkcz"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.874461 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.874483 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.878539 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.883200 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.884255 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-rwd59"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.885080 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-serving-cert\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.889415 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.889463 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.889558 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.890077 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6qgqv"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.895885 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.897614 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wdcs7"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.898289 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.913151 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.913592 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-n5p54"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.915074 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm"] Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.935507 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.949828 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.970875 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 11 08:39:16 crc kubenswrapper[4629]: I1211 08:39:16.990730 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.010004 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.029754 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.065365 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb8zq\" (UniqueName: \"kubernetes.io/projected/bc12104e-ef88-4e7f-aae8-89fdd6641c7e-kube-api-access-jb8zq\") pod \"cluster-image-registry-operator-dc59b4c8b-t7dc6\" (UID: \"bc12104e-ef88-4e7f-aae8-89fdd6641c7e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.092830 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc12104e-ef88-4e7f-aae8-89fdd6641c7e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-t7dc6\" (UID: \"bc12104e-ef88-4e7f-aae8-89fdd6641c7e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.110540 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.130090 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.130103 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.148577 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.148912 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbb99da1-1049-4e03-bc19-49988fc37947-serving-cert\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.149014 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fb614df-6b43-4492-a9ba-7b7eee992120-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sm94r\" (UID: \"9fb614df-6b43-4492-a9ba-7b7eee992120\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.149209 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/449eb3bb-7054-452d-85d6-558aa5efa682-audit-dir\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.149380 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd8h6\" (UniqueName: \"kubernetes.io/projected/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-kube-api-access-zd8h6\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.149577 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.149713 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.149868 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.150054 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9q9w\" (UniqueName: \"kubernetes.io/projected/ec0bf78c-4e35-4129-8976-c8d805991843-kube-api-access-g9q9w\") pod \"route-controller-manager-6576b87f9c-5cc5v\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.150364 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d60662db-4515-4c66-8bfd-e418e6f258fa-config\") pod \"console-operator-58897d9998-8q9kf\" (UID: \"d60662db-4515-4c66-8bfd-e418e6f258fa\") " pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.150464 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-node-pullsecrets\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.150590 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-trusted-ca-bundle\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.150711 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-registry-tls\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.150823 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-registry-certificates\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.150956 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-config\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.150991 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:17.650973002 +0000 UTC m=+145.716388720 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151163 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wvff\" (UniqueName: \"kubernetes.io/projected/dbb99da1-1049-4e03-bc19-49988fc37947-kube-api-access-6wvff\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151490 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca13586-fe9e-43e9-acb8-c3ff573ff1db-config\") pod \"machine-approver-56656f9798-rrmqz\" (UID: \"cca13586-fe9e-43e9-acb8-c3ff573ff1db\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151560 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d60662db-4515-4c66-8bfd-e418e6f258fa-trusted-ca\") pod \"console-operator-58897d9998-8q9kf\" (UID: \"d60662db-4515-4c66-8bfd-e418e6f258fa\") " pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151664 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151722 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-config\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151743 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-image-import-ca\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151757 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-encryption-config\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151793 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c34dd798-968c-4df8-bd03-8308d64fcd8c-serving-cert\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151813 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151829 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-service-ca\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151866 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d957c3d0-bff4-4c92-82ea-3900ea7efa06-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xnsl5\" (UID: \"d957c3d0-bff4-4c92-82ea-3900ea7efa06\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151888 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-oauth-serving-cert\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151948 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85h6k\" (UniqueName: \"kubernetes.io/projected/449eb3bb-7054-452d-85d6-558aa5efa682-kube-api-access-85h6k\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.151988 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c34dd798-968c-4df8-bd03-8308d64fcd8c-encryption-config\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152017 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c34dd798-968c-4df8-bd03-8308d64fcd8c-audit-policies\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152038 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2dzg\" (UniqueName: \"kubernetes.io/projected/c34dd798-968c-4df8-bd03-8308d64fcd8c-kube-api-access-f2dzg\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152059 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-bound-sa-token\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152075 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152094 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2ww4\" (UniqueName: \"kubernetes.io/projected/d60662db-4515-4c66-8bfd-e418e6f258fa-kube-api-access-v2ww4\") pod \"console-operator-58897d9998-8q9kf\" (UID: \"d60662db-4515-4c66-8bfd-e418e6f258fa\") " pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152114 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ec0bf78c-4e35-4129-8976-c8d805991843-client-ca\") pod \"route-controller-manager-6576b87f9c-5cc5v\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152132 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pg48\" (UniqueName: \"kubernetes.io/projected/28a9a7b5-a9ca-4ac4-bef4-da3715a27b67-kube-api-access-6pg48\") pod \"cluster-samples-operator-665b6dd947-p5bwv\" (UID: \"28a9a7b5-a9ca-4ac4-bef4-da3715a27b67\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152154 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152190 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152211 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed370895-f652-499d-8382-69fb05f78f03-config\") pod \"machine-api-operator-5694c8668f-npk4p\" (UID: \"ed370895-f652-499d-8382-69fb05f78f03\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152231 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152248 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb6j5\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-kube-api-access-vb6j5\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152274 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152305 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwbg5\" (UniqueName: \"kubernetes.io/projected/ed370895-f652-499d-8382-69fb05f78f03-kube-api-access-jwbg5\") pod \"machine-api-operator-5694c8668f-npk4p\" (UID: \"ed370895-f652-499d-8382-69fb05f78f03\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152321 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c34dd798-968c-4df8-bd03-8308d64fcd8c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152361 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m975m\" (UniqueName: \"kubernetes.io/projected/9fb614df-6b43-4492-a9ba-7b7eee992120-kube-api-access-m975m\") pod \"openshift-controller-manager-operator-756b6f6bc6-sm94r\" (UID: \"9fb614df-6b43-4492-a9ba-7b7eee992120\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152382 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/28a9a7b5-a9ca-4ac4-bef4-da3715a27b67-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p5bwv\" (UID: \"28a9a7b5-a9ca-4ac4-bef4-da3715a27b67\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152402 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-oauth-config\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152424 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ed370895-f652-499d-8382-69fb05f78f03-images\") pod \"machine-api-operator-5694c8668f-npk4p\" (UID: \"ed370895-f652-499d-8382-69fb05f78f03\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152453 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2kbq\" (UniqueName: \"kubernetes.io/projected/9d819e7a-6f7a-4f6e-bbba-4b334546f172-kube-api-access-f2kbq\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152483 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9nmb\" (UniqueName: \"kubernetes.io/projected/d957c3d0-bff4-4c92-82ea-3900ea7efa06-kube-api-access-w9nmb\") pod \"openshift-config-operator-7777fb866f-xnsl5\" (UID: \"d957c3d0-bff4-4c92-82ea-3900ea7efa06\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152510 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152528 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c34dd798-968c-4df8-bd03-8308d64fcd8c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152577 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ed370895-f652-499d-8382-69fb05f78f03-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-npk4p\" (UID: \"ed370895-f652-499d-8382-69fb05f78f03\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152638 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d60662db-4515-4c66-8bfd-e418e6f258fa-serving-cert\") pod \"console-operator-58897d9998-8q9kf\" (UID: \"d60662db-4515-4c66-8bfd-e418e6f258fa\") " pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152657 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152676 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-etcd-serving-ca\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152717 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c34dd798-968c-4df8-bd03-8308d64fcd8c-audit-dir\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152735 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152751 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-trusted-ca\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152769 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152823 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cca13586-fe9e-43e9-acb8-c3ff573ff1db-machine-approver-tls\") pod \"machine-approver-56656f9798-rrmqz\" (UID: \"cca13586-fe9e-43e9-acb8-c3ff573ff1db\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152839 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzqmt\" (UniqueName: \"kubernetes.io/projected/cca13586-fe9e-43e9-acb8-c3ff573ff1db-kube-api-access-tzqmt\") pod \"machine-approver-56656f9798-rrmqz\" (UID: \"cca13586-fe9e-43e9-acb8-c3ff573ff1db\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152887 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-serving-cert\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152920 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec0bf78c-4e35-4129-8976-c8d805991843-config\") pod \"route-controller-manager-6576b87f9c-5cc5v\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152956 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d957c3d0-bff4-4c92-82ea-3900ea7efa06-serving-cert\") pod \"openshift-config-operator-7777fb866f-xnsl5\" (UID: \"d957c3d0-bff4-4c92-82ea-3900ea7efa06\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.152982 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-audit-policies\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.153002 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-audit\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.153080 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-etcd-client\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.153125 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.153157 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fb614df-6b43-4492-a9ba-7b7eee992120-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sm94r\" (UID: \"9fb614df-6b43-4492-a9ba-7b7eee992120\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.153176 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-audit-dir\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.153199 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-config\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.153222 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.153239 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-client-ca\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.153254 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cca13586-fe9e-43e9-acb8-c3ff573ff1db-auth-proxy-config\") pod \"machine-approver-56656f9798-rrmqz\" (UID: \"cca13586-fe9e-43e9-acb8-c3ff573ff1db\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.153278 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-serving-cert\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.153302 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec0bf78c-4e35-4129-8976-c8d805991843-serving-cert\") pod \"route-controller-manager-6576b87f9c-5cc5v\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.153322 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c34dd798-968c-4df8-bd03-8308d64fcd8c-etcd-client\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.169437 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.189635 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.210503 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.230572 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.252536 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254194 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254375 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d957c3d0-bff4-4c92-82ea-3900ea7efa06-serving-cert\") pod \"openshift-config-operator-7777fb866f-xnsl5\" (UID: \"d957c3d0-bff4-4c92-82ea-3900ea7efa06\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254402 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e92bcf9-0313-43ca-b2ae-fbdd9586aea2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rw8lc\" (UID: \"9e92bcf9-0313-43ca-b2ae-fbdd9586aea2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254434 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-etcd-client\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254451 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fb614df-6b43-4492-a9ba-7b7eee992120-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sm94r\" (UID: \"9fb614df-6b43-4492-a9ba-7b7eee992120\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254468 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-config\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254485 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkqmk\" (UniqueName: \"kubernetes.io/projected/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-kube-api-access-tkqmk\") pod \"collect-profiles-29424030-vtnxm\" (UID: \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254503 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d86l\" (UniqueName: \"kubernetes.io/projected/92906dea-785c-45c4-95fc-9e6ed15fc609-kube-api-access-7d86l\") pod \"ingress-canary-rzxln\" (UID: \"92906dea-785c-45c4-95fc-9e6ed15fc609\") " pod="openshift-ingress-canary/ingress-canary-rzxln" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254538 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254556 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cca13586-fe9e-43e9-acb8-c3ff573ff1db-auth-proxy-config\") pod \"machine-approver-56656f9798-rrmqz\" (UID: \"cca13586-fe9e-43e9-acb8-c3ff573ff1db\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254575 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c34dd798-968c-4df8-bd03-8308d64fcd8c-etcd-client\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254593 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv9cf\" (UniqueName: \"kubernetes.io/projected/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-kube-api-access-xv9cf\") pod \"marketplace-operator-79b997595-rzqjs\" (UID: \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\") " pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254611 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254628 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbb99da1-1049-4e03-bc19-49988fc37947-serving-cert\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254646 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e84237c3-8fcc-4a25-9d54-2422dcc628c5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8v64\" (UID: \"e84237c3-8fcc-4a25-9d54-2422dcc628c5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254662 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ec8356f9-5138-4fe5-867d-22d4d636c0d9-etcd-client\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254682 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fb614df-6b43-4492-a9ba-7b7eee992120-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sm94r\" (UID: \"9fb614df-6b43-4492-a9ba-7b7eee992120\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254699 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-plugins-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254722 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254737 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-registry-tls\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254756 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a33a0903-3d69-4017-b274-4f188c6b89c8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pvsn8\" (UID: \"a33a0903-3d69-4017-b274-4f188c6b89c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254782 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-config\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254797 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-socket-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254814 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0463664b-27fa-42ed-8369-442d93fbdcb5-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8rlll\" (UID: \"0463664b-27fa-42ed-8369-442d93fbdcb5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254833 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-registry-certificates\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254877 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wvff\" (UniqueName: \"kubernetes.io/projected/dbb99da1-1049-4e03-bc19-49988fc37947-kube-api-access-6wvff\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254894 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-csi-data-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254910 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/695005e9-0af9-444d-8117-823af0b1ed46-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mbgh9\" (UID: \"695005e9-0af9-444d-8117-823af0b1ed46\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254934 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.254952 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2lmf\" (UniqueName: \"kubernetes.io/projected/a33a0903-3d69-4017-b274-4f188c6b89c8-kube-api-access-b2lmf\") pod \"ingress-operator-5b745b69d9-pvsn8\" (UID: \"a33a0903-3d69-4017-b274-4f188c6b89c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.255219 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz7w8\" (UniqueName: \"kubernetes.io/projected/9311bda6-d765-4970-9749-3fc2b826b0a3-kube-api-access-hz7w8\") pod \"dns-operator-744455d44c-5ccrx\" (UID: \"9311bda6-d765-4970-9749-3fc2b826b0a3\") " pod="openshift-dns-operator/dns-operator-744455d44c-5ccrx" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.255239 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg4mm\" (UniqueName: \"kubernetes.io/projected/ec8356f9-5138-4fe5-867d-22d4d636c0d9-kube-api-access-gg4mm\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.255256 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c34dd798-968c-4df8-bd03-8308d64fcd8c-serving-cert\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.255273 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.255289 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-mountpoint-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.255448 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fmmgm\" (UID: \"3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fmmgm" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.255495 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.256224 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cca13586-fe9e-43e9-acb8-c3ff573ff1db-auth-proxy-config\") pod \"machine-approver-56656f9798-rrmqz\" (UID: \"cca13586-fe9e-43e9-acb8-c3ff573ff1db\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.256870 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:17.756835725 +0000 UTC m=+145.822251333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258161 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-config\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258193 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fb614df-6b43-4492-a9ba-7b7eee992120-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sm94r\" (UID: \"9fb614df-6b43-4492-a9ba-7b7eee992120\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258217 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c34dd798-968c-4df8-bd03-8308d64fcd8c-encryption-config\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258261 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8bf66625-0e08-4c99-9633-85558c1ce12f-images\") pod \"machine-config-operator-74547568cd-g5p2r\" (UID: \"8bf66625-0e08-4c99-9633-85558c1ce12f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258281 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55-config\") pod \"kube-controller-manager-operator-78b949d7b-lr8q8\" (UID: \"98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258304 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx4fp\" (UniqueName: \"kubernetes.io/projected/aa821340-be6f-4719-8379-c4e6eeda38aa-kube-api-access-nx4fp\") pod \"service-ca-operator-777779d784-h5fwd\" (UID: \"aa821340-be6f-4719-8379-c4e6eeda38aa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258321 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-secret-volume\") pod \"collect-profiles-29424030-vtnxm\" (UID: \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258352 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c34dd798-968c-4df8-bd03-8308d64fcd8c-audit-policies\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258414 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-bound-sa-token\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258444 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pg48\" (UniqueName: \"kubernetes.io/projected/28a9a7b5-a9ca-4ac4-bef4-da3715a27b67-kube-api-access-6pg48\") pod \"cluster-samples-operator-665b6dd947-p5bwv\" (UID: \"28a9a7b5-a9ca-4ac4-bef4-da3715a27b67\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258469 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc96z\" (UniqueName: \"kubernetes.io/projected/695005e9-0af9-444d-8117-823af0b1ed46-kube-api-access-lc96z\") pod \"machine-config-controller-84d6567774-mbgh9\" (UID: \"695005e9-0af9-444d-8117-823af0b1ed46\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258490 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/07febd9a-7813-4722-af4d-9b165be26662-signing-cabundle\") pod \"service-ca-9c57cc56f-6qgqv\" (UID: \"07febd9a-7813-4722-af4d-9b165be26662\") " pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258507 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258524 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb6j5\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-kube-api-access-vb6j5\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258543 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a33a0903-3d69-4017-b274-4f188c6b89c8-metrics-tls\") pod \"ingress-operator-5b745b69d9-pvsn8\" (UID: \"a33a0903-3d69-4017-b274-4f188c6b89c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258560 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8bf66625-0e08-4c99-9633-85558c1ce12f-proxy-tls\") pod \"machine-config-operator-74547568cd-g5p2r\" (UID: \"8bf66625-0e08-4c99-9633-85558c1ce12f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258579 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/92906dea-785c-45c4-95fc-9e6ed15fc609-cert\") pod \"ingress-canary-rzxln\" (UID: \"92906dea-785c-45c4-95fc-9e6ed15fc609\") " pod="openshift-ingress-canary/ingress-canary-rzxln" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258605 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258681 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rzqjs\" (UID: \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\") " pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258698 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-registry-certificates\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258710 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3efc8f47-f897-4072-b1ea-da866707146c-srv-cert\") pod \"catalog-operator-68c6474976-x9fpk\" (UID: \"3efc8f47-f897-4072-b1ea-da866707146c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258728 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/07febd9a-7813-4722-af4d-9b165be26662-signing-key\") pod \"service-ca-9c57cc56f-6qgqv\" (UID: \"07febd9a-7813-4722-af4d-9b165be26662\") " pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258743 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec8356f9-5138-4fe5-867d-22d4d636c0d9-serving-cert\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258768 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ed370895-f652-499d-8382-69fb05f78f03-images\") pod \"machine-api-operator-5694c8668f-npk4p\" (UID: \"ed370895-f652-499d-8382-69fb05f78f03\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258789 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/28a9a7b5-a9ca-4ac4-bef4-da3715a27b67-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p5bwv\" (UID: \"28a9a7b5-a9ca-4ac4-bef4-da3715a27b67\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258811 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-oauth-config\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258832 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2kbq\" (UniqueName: \"kubernetes.io/projected/9d819e7a-6f7a-4f6e-bbba-4b334546f172-kube-api-access-f2kbq\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258868 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb-webhook-cert\") pod \"packageserver-d55dfcdfc-7gkf5\" (UID: \"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258888 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c34dd798-968c-4df8-bd03-8308d64fcd8c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258905 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db4ss\" (UniqueName: \"kubernetes.io/projected/aedfd32e-f597-4753-8fbc-824b562b9c77-kube-api-access-db4ss\") pod \"control-plane-machine-set-operator-78cbb6b69f-m9222\" (UID: \"aedfd32e-f597-4753-8fbc-824b562b9c77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258922 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z6bv\" (UniqueName: \"kubernetes.io/projected/3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8-kube-api-access-4z6bv\") pod \"multus-admission-controller-857f4d67dd-fmmgm\" (UID: \"3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fmmgm" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258940 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.258958 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e84237c3-8fcc-4a25-9d54-2422dcc628c5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8v64\" (UID: \"e84237c3-8fcc-4a25-9d54-2422dcc628c5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259292 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/beac9023-3506-4a56-9ac8-667172e2760d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-v94fb\" (UID: \"beac9023-3506-4a56-9ac8-667172e2760d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259308 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/641fc441-d314-4b07-9fdf-98a81d21744f-metrics-certs\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259329 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259441 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2d4978b-296a-4177-955e-d367860c19ab-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-947kd\" (UID: \"d2d4978b-296a-4177-955e-d367860c19ab\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259468 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cca13586-fe9e-43e9-acb8-c3ff573ff1db-machine-approver-tls\") pod \"machine-approver-56656f9798-rrmqz\" (UID: \"cca13586-fe9e-43e9-acb8-c3ff573ff1db\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259487 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbrrd\" (UniqueName: \"kubernetes.io/projected/ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb-kube-api-access-nbrrd\") pod \"packageserver-d55dfcdfc-7gkf5\" (UID: \"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259509 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-trusted-ca\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259535 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec0bf78c-4e35-4129-8976-c8d805991843-config\") pod \"route-controller-manager-6576b87f9c-5cc5v\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259555 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9311bda6-d765-4970-9749-3fc2b826b0a3-metrics-tls\") pod \"dns-operator-744455d44c-5ccrx\" (UID: \"9311bda6-d765-4970-9749-3fc2b826b0a3\") " pod="openshift-dns-operator/dns-operator-744455d44c-5ccrx" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259579 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259600 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-audit-policies\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259619 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-audit\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259640 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa821340-be6f-4719-8379-c4e6eeda38aa-serving-cert\") pod \"service-ca-operator-777779d784-h5fwd\" (UID: \"aa821340-be6f-4719-8379-c4e6eeda38aa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259661 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gggk9\" (UniqueName: \"kubernetes.io/projected/0463664b-27fa-42ed-8369-442d93fbdcb5-kube-api-access-gggk9\") pod \"package-server-manager-789f6589d5-8rlll\" (UID: \"0463664b-27fa-42ed-8369-442d93fbdcb5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259688 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzqql\" (UniqueName: \"kubernetes.io/projected/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-kube-api-access-bzqql\") pod \"dns-default-n5p54\" (UID: \"6dcbb152-ea3e-4db5-aa09-d60e3768a99c\") " pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259695 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259710 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lr8q8\" (UID: \"98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259735 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-audit-dir\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259759 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ec8356f9-5138-4fe5-867d-22d4d636c0d9-etcd-service-ca\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259784 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-serving-cert\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259805 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-registration-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259831 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-client-ca\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259879 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec0bf78c-4e35-4129-8976-c8d805991843-serving-cert\") pod \"route-controller-manager-6576b87f9c-5cc5v\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259907 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2d4978b-296a-4177-955e-d367860c19ab-config\") pod \"kube-apiserver-operator-766d6c64bb-947kd\" (UID: \"d2d4978b-296a-4177-955e-d367860c19ab\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259929 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/beac9023-3506-4a56-9ac8-667172e2760d-srv-cert\") pod \"olm-operator-6b444d44fb-v94fb\" (UID: \"beac9023-3506-4a56-9ac8-667172e2760d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259964 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd8h6\" (UniqueName: \"kubernetes.io/projected/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-kube-api-access-zd8h6\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.259987 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6v2n\" (UniqueName: \"kubernetes.io/projected/d44b0b79-360a-4d50-abb1-ef1ded04c565-kube-api-access-w6v2n\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.260009 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/449eb3bb-7054-452d-85d6-558aa5efa682-audit-dir\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.260275 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c34dd798-968c-4df8-bd03-8308d64fcd8c-audit-policies\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.262015 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-config\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.265186 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-registry-tls\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.266589 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.266611 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d957c3d0-bff4-4c92-82ea-3900ea7efa06-serving-cert\") pod \"openshift-config-operator-7777fb866f-xnsl5\" (UID: \"d957c3d0-bff4-4c92-82ea-3900ea7efa06\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.266655 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fb614df-6b43-4492-a9ba-7b7eee992120-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sm94r\" (UID: \"9fb614df-6b43-4492-a9ba-7b7eee992120\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.266662 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c34dd798-968c-4df8-bd03-8308d64fcd8c-serving-cert\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.266923 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ed370895-f652-499d-8382-69fb05f78f03-images\") pod \"machine-api-operator-5694c8668f-npk4p\" (UID: \"ed370895-f652-499d-8382-69fb05f78f03\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.267021 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.267503 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c34dd798-968c-4df8-bd03-8308d64fcd8c-etcd-client\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.268101 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c34dd798-968c-4df8-bd03-8308d64fcd8c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.268695 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.269722 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c34dd798-968c-4df8-bd03-8308d64fcd8c-encryption-config\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.269766 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.270205 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.271529 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.271579 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-audit-dir\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.271617 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9q9w\" (UniqueName: \"kubernetes.io/projected/ec0bf78c-4e35-4129-8976-c8d805991843-kube-api-access-g9q9w\") pod \"route-controller-manager-6576b87f9c-5cc5v\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.271642 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d2d4978b-296a-4177-955e-d367860c19ab-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-947kd\" (UID: \"d2d4978b-296a-4177-955e-d367860c19ab\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.271663 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcdst\" (UniqueName: \"kubernetes.io/projected/9e92bcf9-0313-43ca-b2ae-fbdd9586aea2-kube-api-access-dcdst\") pod \"kube-storage-version-migrator-operator-b67b599dd-rw8lc\" (UID: \"9e92bcf9-0313-43ca-b2ae-fbdd9586aea2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.271686 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d60662db-4515-4c66-8bfd-e418e6f258fa-config\") pod \"console-operator-58897d9998-8q9kf\" (UID: \"d60662db-4515-4c66-8bfd-e418e6f258fa\") " pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.271705 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-node-pullsecrets\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.271724 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-trusted-ca-bundle\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.271742 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-config-volume\") pod \"collect-profiles-29424030-vtnxm\" (UID: \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.271762 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/af3b37d7-7663-494c-a57a-419db7475661-node-bootstrap-token\") pod \"machine-config-server-rwd59\" (UID: \"af3b37d7-7663-494c-a57a-419db7475661\") " pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.271800 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhbv9\" (UniqueName: \"kubernetes.io/projected/bfbd1947-e471-47dd-8c36-916c28cfaf79-kube-api-access-dhbv9\") pod \"migrator-59844c95c7-mbkcz\" (UID: \"bfbd1947-e471-47dd-8c36-916c28cfaf79\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mbkcz" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.271865 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/449eb3bb-7054-452d-85d6-558aa5efa682-audit-dir\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.272998 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-audit-policies\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.273318 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec0bf78c-4e35-4129-8976-c8d805991843-config\") pod \"route-controller-manager-6576b87f9c-5cc5v\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.273422 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-trusted-ca\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.273722 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-audit\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.273842 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec0bf78c-4e35-4129-8976-c8d805991843-serving-cert\") pod \"route-controller-manager-6576b87f9c-5cc5v\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.273969 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-node-pullsecrets\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.274231 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twtj4\" (UniqueName: \"kubernetes.io/projected/07febd9a-7813-4722-af4d-9b165be26662-kube-api-access-twtj4\") pod \"service-ca-9c57cc56f-6qgqv\" (UID: \"07febd9a-7813-4722-af4d-9b165be26662\") " pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.274264 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/641fc441-d314-4b07-9fdf-98a81d21744f-stats-auth\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.275116 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-trusted-ca-bundle\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.275719 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d60662db-4515-4c66-8bfd-e418e6f258fa-config\") pod \"console-operator-58897d9998-8q9kf\" (UID: \"d60662db-4515-4c66-8bfd-e418e6f258fa\") " pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276538 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca13586-fe9e-43e9-acb8-c3ff573ff1db-config\") pod \"machine-approver-56656f9798-rrmqz\" (UID: \"cca13586-fe9e-43e9-acb8-c3ff573ff1db\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276564 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d60662db-4515-4c66-8bfd-e418e6f258fa-trusted-ca\") pod \"console-operator-58897d9998-8q9kf\" (UID: \"d60662db-4515-4c66-8bfd-e418e6f258fa\") " pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276588 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-config\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276609 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-image-import-ca\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276628 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-encryption-config\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276649 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfqpc\" (UniqueName: \"kubernetes.io/projected/3efc8f47-f897-4072-b1ea-da866707146c-kube-api-access-nfqpc\") pod \"catalog-operator-68c6474976-x9fpk\" (UID: \"3efc8f47-f897-4072-b1ea-da866707146c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276668 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-config-volume\") pod \"dns-default-n5p54\" (UID: \"6dcbb152-ea3e-4db5-aa09-d60e3768a99c\") " pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276685 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3efc8f47-f897-4072-b1ea-da866707146c-profile-collector-cert\") pod \"catalog-operator-68c6474976-x9fpk\" (UID: \"3efc8f47-f897-4072-b1ea-da866707146c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276706 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d957c3d0-bff4-4c92-82ea-3900ea7efa06-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xnsl5\" (UID: \"d957c3d0-bff4-4c92-82ea-3900ea7efa06\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276723 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh7kz\" (UniqueName: \"kubernetes.io/projected/641fc441-d314-4b07-9fdf-98a81d21744f-kube-api-access-dh7kz\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276743 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-service-ca\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276762 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-oauth-serving-cert\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276779 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/641fc441-d314-4b07-9fdf-98a81d21744f-service-ca-bundle\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276797 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85h6k\" (UniqueName: \"kubernetes.io/projected/449eb3bb-7054-452d-85d6-558aa5efa682-kube-api-access-85h6k\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276814 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec8356f9-5138-4fe5-867d-22d4d636c0d9-config\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276831 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lr8q8\" (UID: \"98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276898 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/aedfd32e-f597-4753-8fbc-824b562b9c77-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-m9222\" (UID: \"aedfd32e-f597-4753-8fbc-824b562b9c77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276919 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ec8356f9-5138-4fe5-867d-22d4d636c0d9-etcd-ca\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276948 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2dzg\" (UniqueName: \"kubernetes.io/projected/c34dd798-968c-4df8-bd03-8308d64fcd8c-kube-api-access-f2dzg\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276969 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.276987 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2ww4\" (UniqueName: \"kubernetes.io/projected/d60662db-4515-4c66-8bfd-e418e6f258fa-kube-api-access-v2ww4\") pod \"console-operator-58897d9998-8q9kf\" (UID: \"d60662db-4515-4c66-8bfd-e418e6f258fa\") " pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277007 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ec0bf78c-4e35-4129-8976-c8d805991843-client-ca\") pod \"route-controller-manager-6576b87f9c-5cc5v\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277024 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa821340-be6f-4719-8379-c4e6eeda38aa-config\") pod \"service-ca-operator-777779d784-h5fwd\" (UID: \"aa821340-be6f-4719-8379-c4e6eeda38aa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277045 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277062 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rzqjs\" (UID: \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\") " pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277086 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttxmt\" (UniqueName: \"kubernetes.io/projected/beac9023-3506-4a56-9ac8-667172e2760d-kube-api-access-ttxmt\") pod \"olm-operator-6b444d44fb-v94fb\" (UID: \"beac9023-3506-4a56-9ac8-667172e2760d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277102 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb-apiservice-cert\") pod \"packageserver-d55dfcdfc-7gkf5\" (UID: \"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277120 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkkmx\" (UniqueName: \"kubernetes.io/projected/83f7a7a2-12d8-43b0-9a3a-3926d3f0092e-kube-api-access-vkkmx\") pod \"downloads-7954f5f757-wbrg7\" (UID: \"83f7a7a2-12d8-43b0-9a3a-3926d3f0092e\") " pod="openshift-console/downloads-7954f5f757-wbrg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277140 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277159 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/af3b37d7-7663-494c-a57a-419db7475661-certs\") pod \"machine-config-server-rwd59\" (UID: \"af3b37d7-7663-494c-a57a-419db7475661\") " pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277181 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-client-ca\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277187 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed370895-f652-499d-8382-69fb05f78f03-config\") pod \"machine-api-operator-5694c8668f-npk4p\" (UID: \"ed370895-f652-499d-8382-69fb05f78f03\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277256 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb-tmpfs\") pod \"packageserver-d55dfcdfc-7gkf5\" (UID: \"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277289 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/695005e9-0af9-444d-8117-823af0b1ed46-proxy-tls\") pod \"machine-config-controller-84d6567774-mbgh9\" (UID: \"695005e9-0af9-444d-8117-823af0b1ed46\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277325 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwbg5\" (UniqueName: \"kubernetes.io/projected/ed370895-f652-499d-8382-69fb05f78f03-kube-api-access-jwbg5\") pod \"machine-api-operator-5694c8668f-npk4p\" (UID: \"ed370895-f652-499d-8382-69fb05f78f03\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277347 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c34dd798-968c-4df8-bd03-8308d64fcd8c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277367 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/641fc441-d314-4b07-9fdf-98a81d21744f-default-certificate\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277396 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9nmb\" (UniqueName: \"kubernetes.io/projected/d957c3d0-bff4-4c92-82ea-3900ea7efa06-kube-api-access-w9nmb\") pod \"openshift-config-operator-7777fb866f-xnsl5\" (UID: \"d957c3d0-bff4-4c92-82ea-3900ea7efa06\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277418 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8bf66625-0e08-4c99-9633-85558c1ce12f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-g5p2r\" (UID: \"8bf66625-0e08-4c99-9633-85558c1ce12f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277443 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m975m\" (UniqueName: \"kubernetes.io/projected/9fb614df-6b43-4492-a9ba-7b7eee992120-kube-api-access-m975m\") pod \"openshift-controller-manager-operator-756b6f6bc6-sm94r\" (UID: \"9fb614df-6b43-4492-a9ba-7b7eee992120\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277466 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e84237c3-8fcc-4a25-9d54-2422dcc628c5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8v64\" (UID: \"e84237c3-8fcc-4a25-9d54-2422dcc628c5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277488 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74rkg\" (UniqueName: \"kubernetes.io/projected/af3b37d7-7663-494c-a57a-419db7475661-kube-api-access-74rkg\") pod \"machine-config-server-rwd59\" (UID: \"af3b37d7-7663-494c-a57a-419db7475661\") " pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277513 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ed370895-f652-499d-8382-69fb05f78f03-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-npk4p\" (UID: \"ed370895-f652-499d-8382-69fb05f78f03\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277533 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d60662db-4515-4c66-8bfd-e418e6f258fa-serving-cert\") pod \"console-operator-58897d9998-8q9kf\" (UID: \"d60662db-4515-4c66-8bfd-e418e6f258fa\") " pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277552 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlk28\" (UniqueName: \"kubernetes.io/projected/8bf66625-0e08-4c99-9633-85558c1ce12f-kube-api-access-zlk28\") pod \"machine-config-operator-74547568cd-g5p2r\" (UID: \"8bf66625-0e08-4c99-9633-85558c1ce12f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277579 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c34dd798-968c-4df8-bd03-8308d64fcd8c-audit-dir\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277598 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-metrics-tls\") pod \"dns-default-n5p54\" (UID: \"6dcbb152-ea3e-4db5-aa09-d60e3768a99c\") " pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277616 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e92bcf9-0313-43ca-b2ae-fbdd9586aea2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rw8lc\" (UID: \"9e92bcf9-0313-43ca-b2ae-fbdd9586aea2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277637 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-etcd-serving-ca\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277646 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca13586-fe9e-43e9-acb8-c3ff573ff1db-config\") pod \"machine-approver-56656f9798-rrmqz\" (UID: \"cca13586-fe9e-43e9-acb8-c3ff573ff1db\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277657 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277678 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a33a0903-3d69-4017-b274-4f188c6b89c8-trusted-ca\") pod \"ingress-operator-5b745b69d9-pvsn8\" (UID: \"a33a0903-3d69-4017-b274-4f188c6b89c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277729 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277750 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzqmt\" (UniqueName: \"kubernetes.io/projected/cca13586-fe9e-43e9-acb8-c3ff573ff1db-kube-api-access-tzqmt\") pod \"machine-approver-56656f9798-rrmqz\" (UID: \"cca13586-fe9e-43e9-acb8-c3ff573ff1db\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.277773 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-serving-cert\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.278020 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d60662db-4515-4c66-8bfd-e418e6f258fa-trusted-ca\") pod \"console-operator-58897d9998-8q9kf\" (UID: \"d60662db-4515-4c66-8bfd-e418e6f258fa\") " pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.278096 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c34dd798-968c-4df8-bd03-8308d64fcd8c-audit-dir\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.278720 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ec0bf78c-4e35-4129-8976-c8d805991843-client-ca\") pod \"route-controller-manager-6576b87f9c-5cc5v\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.278757 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-etcd-serving-ca\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.279198 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-config\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.279253 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c34dd798-968c-4df8-bd03-8308d64fcd8c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.279749 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.279906 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-service-ca\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.280064 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-image-import-ca\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.280480 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-serving-cert\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.280595 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed370895-f652-499d-8382-69fb05f78f03-config\") pod \"machine-api-operator-5694c8668f-npk4p\" (UID: \"ed370895-f652-499d-8382-69fb05f78f03\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.281769 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-oauth-serving-cert\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.285992 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cca13586-fe9e-43e9-acb8-c3ff573ff1db-machine-approver-tls\") pod \"machine-approver-56656f9798-rrmqz\" (UID: \"cca13586-fe9e-43e9-acb8-c3ff573ff1db\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.286988 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.289562 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d957c3d0-bff4-4c92-82ea-3900ea7efa06-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xnsl5\" (UID: \"d957c3d0-bff4-4c92-82ea-3900ea7efa06\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.289820 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.289926 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-serving-cert\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.291125 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbb99da1-1049-4e03-bc19-49988fc37947-serving-cert\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.291459 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.293054 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d60662db-4515-4c66-8bfd-e418e6f258fa-serving-cert\") pod \"console-operator-58897d9998-8q9kf\" (UID: \"d60662db-4515-4c66-8bfd-e418e6f258fa\") " pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.293194 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.294295 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.295661 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.296050 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.296160 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-etcd-client\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.296538 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-oauth-config\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.296766 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.296794 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/28a9a7b5-a9ca-4ac4-bef4-da3715a27b67-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p5bwv\" (UID: \"28a9a7b5-a9ca-4ac4-bef4-da3715a27b67\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.299066 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-encryption-config\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.302016 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ed370895-f652-499d-8382-69fb05f78f03-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-npk4p\" (UID: \"ed370895-f652-499d-8382-69fb05f78f03\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.313313 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.333100 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.352660 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.352702 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6"] Dec 11 08:39:17 crc kubenswrapper[4629]: W1211 08:39:17.362517 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc12104e_ef88_4e7f_aae8_89fdd6641c7e.slice/crio-1fbfe40fb4ec33adad21d1a7247b370d9fee2ca4b933b746b956c723e75c7285 WatchSource:0}: Error finding container 1fbfe40fb4ec33adad21d1a7247b370d9fee2ca4b933b746b956c723e75c7285: Status 404 returned error can't find the container with id 1fbfe40fb4ec33adad21d1a7247b370d9fee2ca4b933b746b956c723e75c7285 Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.369673 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.379730 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-socket-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.379799 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0463664b-27fa-42ed-8369-442d93fbdcb5-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8rlll\" (UID: \"0463664b-27fa-42ed-8369-442d93fbdcb5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.379833 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-csi-data-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.379908 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/695005e9-0af9-444d-8117-823af0b1ed46-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mbgh9\" (UID: \"695005e9-0af9-444d-8117-823af0b1ed46\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380009 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz7w8\" (UniqueName: \"kubernetes.io/projected/9311bda6-d765-4970-9749-3fc2b826b0a3-kube-api-access-hz7w8\") pod \"dns-operator-744455d44c-5ccrx\" (UID: \"9311bda6-d765-4970-9749-3fc2b826b0a3\") " pod="openshift-dns-operator/dns-operator-744455d44c-5ccrx" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380057 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2lmf\" (UniqueName: \"kubernetes.io/projected/a33a0903-3d69-4017-b274-4f188c6b89c8-kube-api-access-b2lmf\") pod \"ingress-operator-5b745b69d9-pvsn8\" (UID: \"a33a0903-3d69-4017-b274-4f188c6b89c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380082 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg4mm\" (UniqueName: \"kubernetes.io/projected/ec8356f9-5138-4fe5-867d-22d4d636c0d9-kube-api-access-gg4mm\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380108 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-mountpoint-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380129 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fmmgm\" (UID: \"3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fmmgm" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380151 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8bf66625-0e08-4c99-9633-85558c1ce12f-images\") pod \"machine-config-operator-74547568cd-g5p2r\" (UID: \"8bf66625-0e08-4c99-9633-85558c1ce12f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380174 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55-config\") pod \"kube-controller-manager-operator-78b949d7b-lr8q8\" (UID: \"98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380205 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx4fp\" (UniqueName: \"kubernetes.io/projected/aa821340-be6f-4719-8379-c4e6eeda38aa-kube-api-access-nx4fp\") pod \"service-ca-operator-777779d784-h5fwd\" (UID: \"aa821340-be6f-4719-8379-c4e6eeda38aa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380228 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-secret-volume\") pod \"collect-profiles-29424030-vtnxm\" (UID: \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380272 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/07febd9a-7813-4722-af4d-9b165be26662-signing-cabundle\") pod \"service-ca-9c57cc56f-6qgqv\" (UID: \"07febd9a-7813-4722-af4d-9b165be26662\") " pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380295 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc96z\" (UniqueName: \"kubernetes.io/projected/695005e9-0af9-444d-8117-823af0b1ed46-kube-api-access-lc96z\") pod \"machine-config-controller-84d6567774-mbgh9\" (UID: \"695005e9-0af9-444d-8117-823af0b1ed46\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380318 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8bf66625-0e08-4c99-9633-85558c1ce12f-proxy-tls\") pod \"machine-config-operator-74547568cd-g5p2r\" (UID: \"8bf66625-0e08-4c99-9633-85558c1ce12f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380345 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a33a0903-3d69-4017-b274-4f188c6b89c8-metrics-tls\") pod \"ingress-operator-5b745b69d9-pvsn8\" (UID: \"a33a0903-3d69-4017-b274-4f188c6b89c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380365 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/92906dea-785c-45c4-95fc-9e6ed15fc609-cert\") pod \"ingress-canary-rzxln\" (UID: \"92906dea-785c-45c4-95fc-9e6ed15fc609\") " pod="openshift-ingress-canary/ingress-canary-rzxln" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380399 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3efc8f47-f897-4072-b1ea-da866707146c-srv-cert\") pod \"catalog-operator-68c6474976-x9fpk\" (UID: \"3efc8f47-f897-4072-b1ea-da866707146c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380446 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rzqjs\" (UID: \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\") " pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380510 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/07febd9a-7813-4722-af4d-9b165be26662-signing-key\") pod \"service-ca-9c57cc56f-6qgqv\" (UID: \"07febd9a-7813-4722-af4d-9b165be26662\") " pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380549 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec8356f9-5138-4fe5-867d-22d4d636c0d9-serving-cert\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380605 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb-webhook-cert\") pod \"packageserver-d55dfcdfc-7gkf5\" (UID: \"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380651 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db4ss\" (UniqueName: \"kubernetes.io/projected/aedfd32e-f597-4753-8fbc-824b562b9c77-kube-api-access-db4ss\") pod \"control-plane-machine-set-operator-78cbb6b69f-m9222\" (UID: \"aedfd32e-f597-4753-8fbc-824b562b9c77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380699 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z6bv\" (UniqueName: \"kubernetes.io/projected/3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8-kube-api-access-4z6bv\") pod \"multus-admission-controller-857f4d67dd-fmmgm\" (UID: \"3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fmmgm" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380724 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e84237c3-8fcc-4a25-9d54-2422dcc628c5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8v64\" (UID: \"e84237c3-8fcc-4a25-9d54-2422dcc628c5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380748 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/beac9023-3506-4a56-9ac8-667172e2760d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-v94fb\" (UID: \"beac9023-3506-4a56-9ac8-667172e2760d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380767 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/641fc441-d314-4b07-9fdf-98a81d21744f-metrics-certs\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380788 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2d4978b-296a-4177-955e-d367860c19ab-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-947kd\" (UID: \"d2d4978b-296a-4177-955e-d367860c19ab\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380813 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbrrd\" (UniqueName: \"kubernetes.io/projected/ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb-kube-api-access-nbrrd\") pod \"packageserver-d55dfcdfc-7gkf5\" (UID: \"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380835 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9311bda6-d765-4970-9749-3fc2b826b0a3-metrics-tls\") pod \"dns-operator-744455d44c-5ccrx\" (UID: \"9311bda6-d765-4970-9749-3fc2b826b0a3\") " pod="openshift-dns-operator/dns-operator-744455d44c-5ccrx" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380879 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa821340-be6f-4719-8379-c4e6eeda38aa-serving-cert\") pod \"service-ca-operator-777779d784-h5fwd\" (UID: \"aa821340-be6f-4719-8379-c4e6eeda38aa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380902 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gggk9\" (UniqueName: \"kubernetes.io/projected/0463664b-27fa-42ed-8369-442d93fbdcb5-kube-api-access-gggk9\") pod \"package-server-manager-789f6589d5-8rlll\" (UID: \"0463664b-27fa-42ed-8369-442d93fbdcb5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380927 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ec8356f9-5138-4fe5-867d-22d4d636c0d9-etcd-service-ca\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380949 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzqql\" (UniqueName: \"kubernetes.io/projected/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-kube-api-access-bzqql\") pod \"dns-default-n5p54\" (UID: \"6dcbb152-ea3e-4db5-aa09-d60e3768a99c\") " pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380972 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lr8q8\" (UID: \"98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.380999 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-registration-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381018 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/beac9023-3506-4a56-9ac8-667172e2760d-srv-cert\") pod \"olm-operator-6b444d44fb-v94fb\" (UID: \"beac9023-3506-4a56-9ac8-667172e2760d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381056 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2d4978b-296a-4177-955e-d367860c19ab-config\") pod \"kube-apiserver-operator-766d6c64bb-947kd\" (UID: \"d2d4978b-296a-4177-955e-d367860c19ab\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381086 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6v2n\" (UniqueName: \"kubernetes.io/projected/d44b0b79-360a-4d50-abb1-ef1ded04c565-kube-api-access-w6v2n\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381120 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d2d4978b-296a-4177-955e-d367860c19ab-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-947kd\" (UID: \"d2d4978b-296a-4177-955e-d367860c19ab\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381156 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcdst\" (UniqueName: \"kubernetes.io/projected/9e92bcf9-0313-43ca-b2ae-fbdd9586aea2-kube-api-access-dcdst\") pod \"kube-storage-version-migrator-operator-b67b599dd-rw8lc\" (UID: \"9e92bcf9-0313-43ca-b2ae-fbdd9586aea2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381187 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-config-volume\") pod \"collect-profiles-29424030-vtnxm\" (UID: \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381209 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/af3b37d7-7663-494c-a57a-419db7475661-node-bootstrap-token\") pod \"machine-config-server-rwd59\" (UID: \"af3b37d7-7663-494c-a57a-419db7475661\") " pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381243 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhbv9\" (UniqueName: \"kubernetes.io/projected/bfbd1947-e471-47dd-8c36-916c28cfaf79-kube-api-access-dhbv9\") pod \"migrator-59844c95c7-mbkcz\" (UID: \"bfbd1947-e471-47dd-8c36-916c28cfaf79\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mbkcz" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381269 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twtj4\" (UniqueName: \"kubernetes.io/projected/07febd9a-7813-4722-af4d-9b165be26662-kube-api-access-twtj4\") pod \"service-ca-9c57cc56f-6qgqv\" (UID: \"07febd9a-7813-4722-af4d-9b165be26662\") " pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381292 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/641fc441-d314-4b07-9fdf-98a81d21744f-stats-auth\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381323 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfqpc\" (UniqueName: \"kubernetes.io/projected/3efc8f47-f897-4072-b1ea-da866707146c-kube-api-access-nfqpc\") pod \"catalog-operator-68c6474976-x9fpk\" (UID: \"3efc8f47-f897-4072-b1ea-da866707146c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381302 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-socket-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381368 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55-config\") pod \"kube-controller-manager-operator-78b949d7b-lr8q8\" (UID: \"98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381373 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3efc8f47-f897-4072-b1ea-da866707146c-profile-collector-cert\") pod \"catalog-operator-68c6474976-x9fpk\" (UID: \"3efc8f47-f897-4072-b1ea-da866707146c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381421 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-mountpoint-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381424 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-config-volume\") pod \"dns-default-n5p54\" (UID: \"6dcbb152-ea3e-4db5-aa09-d60e3768a99c\") " pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381774 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh7kz\" (UniqueName: \"kubernetes.io/projected/641fc441-d314-4b07-9fdf-98a81d21744f-kube-api-access-dh7kz\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381815 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/641fc441-d314-4b07-9fdf-98a81d21744f-service-ca-bundle\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381855 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec8356f9-5138-4fe5-867d-22d4d636c0d9-config\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381873 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lr8q8\" (UID: \"98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381900 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/aedfd32e-f597-4753-8fbc-824b562b9c77-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-m9222\" (UID: \"aedfd32e-f597-4753-8fbc-824b562b9c77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381920 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ec8356f9-5138-4fe5-867d-22d4d636c0d9-etcd-ca\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381938 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa821340-be6f-4719-8379-c4e6eeda38aa-config\") pod \"service-ca-operator-777779d784-h5fwd\" (UID: \"aa821340-be6f-4719-8379-c4e6eeda38aa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381975 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rzqjs\" (UID: \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\") " pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381996 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttxmt\" (UniqueName: \"kubernetes.io/projected/beac9023-3506-4a56-9ac8-667172e2760d-kube-api-access-ttxmt\") pod \"olm-operator-6b444d44fb-v94fb\" (UID: \"beac9023-3506-4a56-9ac8-667172e2760d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382019 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb-apiservice-cert\") pod \"packageserver-d55dfcdfc-7gkf5\" (UID: \"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382038 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkkmx\" (UniqueName: \"kubernetes.io/projected/83f7a7a2-12d8-43b0-9a3a-3926d3f0092e-kube-api-access-vkkmx\") pod \"downloads-7954f5f757-wbrg7\" (UID: \"83f7a7a2-12d8-43b0-9a3a-3926d3f0092e\") " pod="openshift-console/downloads-7954f5f757-wbrg7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382056 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/af3b37d7-7663-494c-a57a-419db7475661-certs\") pod \"machine-config-server-rwd59\" (UID: \"af3b37d7-7663-494c-a57a-419db7475661\") " pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382074 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb-tmpfs\") pod \"packageserver-d55dfcdfc-7gkf5\" (UID: \"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382087 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/695005e9-0af9-444d-8117-823af0b1ed46-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mbgh9\" (UID: \"695005e9-0af9-444d-8117-823af0b1ed46\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382097 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/695005e9-0af9-444d-8117-823af0b1ed46-proxy-tls\") pod \"machine-config-controller-84d6567774-mbgh9\" (UID: \"695005e9-0af9-444d-8117-823af0b1ed46\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382213 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/641fc441-d314-4b07-9fdf-98a81d21744f-default-certificate\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382258 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8bf66625-0e08-4c99-9633-85558c1ce12f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-g5p2r\" (UID: \"8bf66625-0e08-4c99-9633-85558c1ce12f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382287 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e84237c3-8fcc-4a25-9d54-2422dcc628c5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8v64\" (UID: \"e84237c3-8fcc-4a25-9d54-2422dcc628c5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382318 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74rkg\" (UniqueName: \"kubernetes.io/projected/af3b37d7-7663-494c-a57a-419db7475661-kube-api-access-74rkg\") pod \"machine-config-server-rwd59\" (UID: \"af3b37d7-7663-494c-a57a-419db7475661\") " pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382352 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlk28\" (UniqueName: \"kubernetes.io/projected/8bf66625-0e08-4c99-9633-85558c1ce12f-kube-api-access-zlk28\") pod \"machine-config-operator-74547568cd-g5p2r\" (UID: \"8bf66625-0e08-4c99-9633-85558c1ce12f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382380 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-metrics-tls\") pod \"dns-default-n5p54\" (UID: \"6dcbb152-ea3e-4db5-aa09-d60e3768a99c\") " pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382407 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e92bcf9-0313-43ca-b2ae-fbdd9586aea2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rw8lc\" (UID: \"9e92bcf9-0313-43ca-b2ae-fbdd9586aea2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382433 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a33a0903-3d69-4017-b274-4f188c6b89c8-trusted-ca\") pod \"ingress-operator-5b745b69d9-pvsn8\" (UID: \"a33a0903-3d69-4017-b274-4f188c6b89c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382513 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e92bcf9-0313-43ca-b2ae-fbdd9586aea2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rw8lc\" (UID: \"9e92bcf9-0313-43ca-b2ae-fbdd9586aea2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382541 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d86l\" (UniqueName: \"kubernetes.io/projected/92906dea-785c-45c4-95fc-9e6ed15fc609-kube-api-access-7d86l\") pod \"ingress-canary-rzxln\" (UID: \"92906dea-785c-45c4-95fc-9e6ed15fc609\") " pod="openshift-ingress-canary/ingress-canary-rzxln" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382572 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkqmk\" (UniqueName: \"kubernetes.io/projected/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-kube-api-access-tkqmk\") pod \"collect-profiles-29424030-vtnxm\" (UID: \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382607 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv9cf\" (UniqueName: \"kubernetes.io/projected/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-kube-api-access-xv9cf\") pod \"marketplace-operator-79b997595-rzqjs\" (UID: \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\") " pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382635 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e84237c3-8fcc-4a25-9d54-2422dcc628c5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8v64\" (UID: \"e84237c3-8fcc-4a25-9d54-2422dcc628c5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382659 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ec8356f9-5138-4fe5-867d-22d4d636c0d9-etcd-client\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382687 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382711 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-plugins-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.382745 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a33a0903-3d69-4017-b274-4f188c6b89c8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pvsn8\" (UID: \"a33a0903-3d69-4017-b274-4f188c6b89c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.383014 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/641fc441-d314-4b07-9fdf-98a81d21744f-service-ca-bundle\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.383560 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec8356f9-5138-4fe5-867d-22d4d636c0d9-config\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.383564 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb-tmpfs\") pod \"packageserver-d55dfcdfc-7gkf5\" (UID: \"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.381153 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-csi-data-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.385209 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-registration-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.385245 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ec8356f9-5138-4fe5-867d-22d4d636c0d9-etcd-ca\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.385502 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:17.885489506 +0000 UTC m=+145.950905114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.385688 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8bf66625-0e08-4c99-9633-85558c1ce12f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-g5p2r\" (UID: \"8bf66625-0e08-4c99-9633-85558c1ce12f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.385973 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d44b0b79-360a-4d50-abb1-ef1ded04c565-plugins-dir\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.386069 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2d4978b-296a-4177-955e-d367860c19ab-config\") pod \"kube-apiserver-operator-766d6c64bb-947kd\" (UID: \"d2d4978b-296a-4177-955e-d367860c19ab\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.386751 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ec8356f9-5138-4fe5-867d-22d4d636c0d9-etcd-service-ca\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.386813 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec8356f9-5138-4fe5-867d-22d4d636c0d9-serving-cert\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.390783 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/641fc441-d314-4b07-9fdf-98a81d21744f-stats-auth\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.391244 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/641fc441-d314-4b07-9fdf-98a81d21744f-metrics-certs\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.391295 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.391602 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d2d4978b-296a-4177-955e-d367860c19ab-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-947kd\" (UID: \"d2d4978b-296a-4177-955e-d367860c19ab\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.391783 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/641fc441-d314-4b07-9fdf-98a81d21744f-default-certificate\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.392160 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lr8q8\" (UID: \"98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.392752 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9311bda6-d765-4970-9749-3fc2b826b0a3-metrics-tls\") pod \"dns-operator-744455d44c-5ccrx\" (UID: \"9311bda6-d765-4970-9749-3fc2b826b0a3\") " pod="openshift-dns-operator/dns-operator-744455d44c-5ccrx" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.393442 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ec8356f9-5138-4fe5-867d-22d4d636c0d9-etcd-client\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.411488 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.415901 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a33a0903-3d69-4017-b274-4f188c6b89c8-metrics-tls\") pod \"ingress-operator-5b745b69d9-pvsn8\" (UID: \"a33a0903-3d69-4017-b274-4f188c6b89c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.437514 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.449412 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a33a0903-3d69-4017-b274-4f188c6b89c8-trusted-ca\") pod \"ingress-operator-5b745b69d9-pvsn8\" (UID: \"a33a0903-3d69-4017-b274-4f188c6b89c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.450221 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.470477 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.484123 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.484383 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:17.984337074 +0000 UTC m=+146.049752682 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.484483 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.485013 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:17.985005255 +0000 UTC m=+146.050420863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.490389 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.510569 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.529891 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.550685 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.570105 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.572974 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8bf66625-0e08-4c99-9633-85558c1ce12f-images\") pod \"machine-config-operator-74547568cd-g5p2r\" (UID: \"8bf66625-0e08-4c99-9633-85558c1ce12f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.586400 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.586495 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.086475858 +0000 UTC m=+146.151891466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.586697 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.587153 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.087066167 +0000 UTC m=+146.152481775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.589599 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.610046 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.619525 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e92bcf9-0313-43ca-b2ae-fbdd9586aea2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rw8lc\" (UID: \"9e92bcf9-0313-43ca-b2ae-fbdd9586aea2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.630763 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.636184 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e92bcf9-0313-43ca-b2ae-fbdd9586aea2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rw8lc\" (UID: \"9e92bcf9-0313-43ca-b2ae-fbdd9586aea2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.649928 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.669999 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.679613 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e84237c3-8fcc-4a25-9d54-2422dcc628c5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8v64\" (UID: \"e84237c3-8fcc-4a25-9d54-2422dcc628c5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.690030 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.690191 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.690374 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.190349447 +0000 UTC m=+146.255765055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.690826 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.691473 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.191463682 +0000 UTC m=+146.256879290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.710043 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.715750 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e84237c3-8fcc-4a25-9d54-2422dcc628c5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8v64\" (UID: \"e84237c3-8fcc-4a25-9d54-2422dcc628c5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.730081 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.736272 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/695005e9-0af9-444d-8117-823af0b1ed46-proxy-tls\") pod \"machine-config-controller-84d6567774-mbgh9\" (UID: \"695005e9-0af9-444d-8117-823af0b1ed46\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.750268 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.771137 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.775570 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8bf66625-0e08-4c99-9633-85558c1ce12f-proxy-tls\") pod \"machine-config-operator-74547568cd-g5p2r\" (UID: \"8bf66625-0e08-4c99-9633-85558c1ce12f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.790794 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.792005 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.792189 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.29215931 +0000 UTC m=+146.357574918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.792560 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.792902 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.292890134 +0000 UTC m=+146.358305732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.808728 4629 request.go:700] Waited for 1.005318924s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serviceaccount-dockercfg-rq7zk&limit=500&resourceVersion=0 Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.811787 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.830494 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.838075 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0463664b-27fa-42ed-8369-442d93fbdcb5-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8rlll\" (UID: \"0463664b-27fa-42ed-8369-442d93fbdcb5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.851403 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.870588 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.889596 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.893992 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.894143 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.394118098 +0000 UTC m=+146.459533706 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.894788 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.895136 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.395127411 +0000 UTC m=+146.460543019 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.904374 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" event={"ID":"bc12104e-ef88-4e7f-aae8-89fdd6641c7e","Type":"ContainerStarted","Data":"18a4f5f3792c9da3ee9da8f9969fd7c15e144632d5677061f9b08186564e1336"} Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.904437 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" event={"ID":"bc12104e-ef88-4e7f-aae8-89fdd6641c7e","Type":"ContainerStarted","Data":"1fbfe40fb4ec33adad21d1a7247b370d9fee2ca4b933b746b956c723e75c7285"} Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.910864 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.918651 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa821340-be6f-4719-8379-c4e6eeda38aa-serving-cert\") pod \"service-ca-operator-777779d784-h5fwd\" (UID: \"aa821340-be6f-4719-8379-c4e6eeda38aa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.931198 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.935681 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa821340-be6f-4719-8379-c4e6eeda38aa-config\") pod \"service-ca-operator-777779d784-h5fwd\" (UID: \"aa821340-be6f-4719-8379-c4e6eeda38aa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.951207 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.969811 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.976355 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-config-volume\") pod \"collect-profiles-29424030-vtnxm\" (UID: \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.990482 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.995997 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-secret-volume\") pod \"collect-profiles-29424030-vtnxm\" (UID: \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.996519 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.996724 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.496679685 +0000 UTC m=+146.562095303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.996956 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3efc8f47-f897-4072-b1ea-da866707146c-profile-collector-cert\") pod \"catalog-operator-68c6474976-x9fpk\" (UID: \"3efc8f47-f897-4072-b1ea-da866707146c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:17 crc kubenswrapper[4629]: I1211 08:39:17.997416 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:17 crc kubenswrapper[4629]: E1211 08:39:17.997956 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.497942246 +0000 UTC m=+146.563357854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.000379 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/beac9023-3506-4a56-9ac8-667172e2760d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-v94fb\" (UID: \"beac9023-3506-4a56-9ac8-667172e2760d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.010604 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.030988 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.037936 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/aedfd32e-f597-4753-8fbc-824b562b9c77-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-m9222\" (UID: \"aedfd32e-f597-4753-8fbc-824b562b9c77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.050656 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.070974 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.076739 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb-apiservice-cert\") pod \"packageserver-d55dfcdfc-7gkf5\" (UID: \"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.078166 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb-webhook-cert\") pod \"packageserver-d55dfcdfc-7gkf5\" (UID: \"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.089774 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.096357 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3efc8f47-f897-4072-b1ea-da866707146c-srv-cert\") pod \"catalog-operator-68c6474976-x9fpk\" (UID: \"3efc8f47-f897-4072-b1ea-da866707146c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.098359 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.099974 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.599958106 +0000 UTC m=+146.665373714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.109696 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.130145 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.136124 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fmmgm\" (UID: \"3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fmmgm" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.149442 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.169523 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.176861 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rzqjs\" (UID: \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\") " pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.190133 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.201241 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.201349 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.201433 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.201468 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.201604 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.201873 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.701833911 +0000 UTC m=+146.767249599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.202764 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.204827 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.204915 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.204999 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.210394 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.236940 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.244445 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rzqjs\" (UID: \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\") " pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.269637 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.290329 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.296283 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/07febd9a-7813-4722-af4d-9b165be26662-signing-key\") pod \"service-ca-9c57cc56f-6qgqv\" (UID: \"07febd9a-7813-4722-af4d-9b165be26662\") " pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.302239 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.302366 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.802348583 +0000 UTC m=+146.867764191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.302663 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.303518 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.80349303 +0000 UTC m=+146.868908638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.309543 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.329229 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.349452 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.353993 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/07febd9a-7813-4722-af4d-9b165be26662-signing-cabundle\") pod \"service-ca-9c57cc56f-6qgqv\" (UID: \"07febd9a-7813-4722-af4d-9b165be26662\") " pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.370515 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.382099 4629 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.382124 4629 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.382219 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/92906dea-785c-45c4-95fc-9e6ed15fc609-cert podName:92906dea-785c-45c4-95fc-9e6ed15fc609 nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.882194865 +0000 UTC m=+146.947610473 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/92906dea-785c-45c4-95fc-9e6ed15fc609-cert") pod "ingress-canary-rzxln" (UID: "92906dea-785c-45c4-95fc-9e6ed15fc609") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.382245 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-config-volume podName:6dcbb152-ea3e-4db5-aa09-d60e3768a99c nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.882230746 +0000 UTC m=+146.947646354 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-config-volume") pod "dns-default-n5p54" (UID: "6dcbb152-ea3e-4db5-aa09-d60e3768a99c") : failed to sync configmap cache: timed out waiting for the condition Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.384257 4629 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.384336 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af3b37d7-7663-494c-a57a-419db7475661-certs podName:af3b37d7-7663-494c-a57a-419db7475661 nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.884319302 +0000 UTC m=+146.949734900 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/af3b37d7-7663-494c-a57a-419db7475661-certs") pod "machine-config-server-rwd59" (UID: "af3b37d7-7663-494c-a57a-419db7475661") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.384978 4629 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.385062 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af3b37d7-7663-494c-a57a-419db7475661-node-bootstrap-token podName:af3b37d7-7663-494c-a57a-419db7475661 nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.885046466 +0000 UTC m=+146.950462144 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/af3b37d7-7663-494c-a57a-419db7475661-node-bootstrap-token") pod "machine-config-server-rwd59" (UID: "af3b37d7-7663-494c-a57a-419db7475661") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.385357 4629 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.385433 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/beac9023-3506-4a56-9ac8-667172e2760d-srv-cert podName:beac9023-3506-4a56-9ac8-667172e2760d nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.885418678 +0000 UTC m=+146.950834366 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/beac9023-3506-4a56-9ac8-667172e2760d-srv-cert") pod "olm-operator-6b444d44fb-v94fb" (UID: "beac9023-3506-4a56-9ac8-667172e2760d") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.386708 4629 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.386798 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-metrics-tls podName:6dcbb152-ea3e-4db5-aa09-d60e3768a99c nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.886777571 +0000 UTC m=+146.952193179 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-metrics-tls") pod "dns-default-n5p54" (UID: "6dcbb152-ea3e-4db5-aa09-d60e3768a99c") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.389455 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.405936 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.905919732 +0000 UTC m=+146.971335330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.405963 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.406122 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.406669 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:18.906653656 +0000 UTC m=+146.972069264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.410986 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.414199 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.422718 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.429737 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.430720 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.450292 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.495666 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvp97\" (UniqueName: \"kubernetes.io/projected/9167ddcb-69c2-4861-80e2-1ab4419b8e93-kube-api-access-zvp97\") pod \"openshift-apiserver-operator-796bbdcf4f-bnswd\" (UID: \"9167ddcb-69c2-4861-80e2-1ab4419b8e93\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.507101 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.507371 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:19.007349474 +0000 UTC m=+147.072765092 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.507986 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.508307 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:19.008298634 +0000 UTC m=+147.073714242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.510669 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.515767 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6h74\" (UniqueName: \"kubernetes.io/projected/fdb51e07-8d01-4c36-9e03-68c0dbc82c0a-kube-api-access-b6h74\") pod \"authentication-operator-69f744f599-bktzv\" (UID: \"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.532195 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.551260 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.570003 4629 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.590570 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.609865 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.610976 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:19.110936364 +0000 UTC m=+147.176351972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.611800 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 11 08:39:18 crc kubenswrapper[4629]: W1211 08:39:18.622915 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-64b74c615bc6744d1b80bf778b5d291d73bf1aa81282e348b64b42f60613d1d7 WatchSource:0}: Error finding container 64b74c615bc6744d1b80bf778b5d291d73bf1aa81282e348b64b42f60613d1d7: Status 404 returned error can't find the container with id 64b74c615bc6744d1b80bf778b5d291d73bf1aa81282e348b64b42f60613d1d7 Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.629926 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.651627 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.667149 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.669954 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.677934 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.708644 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wvff\" (UniqueName: \"kubernetes.io/projected/dbb99da1-1049-4e03-bc19-49988fc37947-kube-api-access-6wvff\") pod \"controller-manager-879f6c89f-2c7tp\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.713300 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.714006 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:19.213993857 +0000 UTC m=+147.279409465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.727713 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb6j5\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-kube-api-access-vb6j5\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.751606 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2kbq\" (UniqueName: \"kubernetes.io/projected/9d819e7a-6f7a-4f6e-bbba-4b334546f172-kube-api-access-f2kbq\") pod \"console-f9d7485db-lhszp\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.758548 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.764784 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-bound-sa-token\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.792961 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pg48\" (UniqueName: \"kubernetes.io/projected/28a9a7b5-a9ca-4ac4-bef4-da3715a27b67-kube-api-access-6pg48\") pod \"cluster-samples-operator-665b6dd947-p5bwv\" (UID: \"28a9a7b5-a9ca-4ac4-bef4-da3715a27b67\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.814725 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.815555 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:19.31549408 +0000 UTC m=+147.380909688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.821914 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd8h6\" (UniqueName: \"kubernetes.io/projected/0c0faec1-d9b4-49c4-8595-381e5d0c4a56-kube-api-access-zd8h6\") pod \"apiserver-76f77b778f-t7tg7\" (UID: \"0c0faec1-d9b4-49c4-8595-381e5d0c4a56\") " pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.830173 4629 request.go:700] Waited for 1.552014104s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console-operator/serviceaccounts/console-operator/token Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.840589 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9q9w\" (UniqueName: \"kubernetes.io/projected/ec0bf78c-4e35-4129-8976-c8d805991843-kube-api-access-g9q9w\") pod \"route-controller-manager-6576b87f9c-5cc5v\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.843163 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.865460 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2ww4\" (UniqueName: \"kubernetes.io/projected/d60662db-4515-4c66-8bfd-e418e6f258fa-kube-api-access-v2ww4\") pod \"console-operator-58897d9998-8q9kf\" (UID: \"d60662db-4515-4c66-8bfd-e418e6f258fa\") " pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.868699 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwbg5\" (UniqueName: \"kubernetes.io/projected/ed370895-f652-499d-8382-69fb05f78f03-kube-api-access-jwbg5\") pod \"machine-api-operator-5694c8668f-npk4p\" (UID: \"ed370895-f652-499d-8382-69fb05f78f03\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.891457 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9nmb\" (UniqueName: \"kubernetes.io/projected/d957c3d0-bff4-4c92-82ea-3900ea7efa06-kube-api-access-w9nmb\") pod \"openshift-config-operator-7777fb866f-xnsl5\" (UID: \"d957c3d0-bff4-4c92-82ea-3900ea7efa06\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.915933 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m975m\" (UniqueName: \"kubernetes.io/projected/9fb614df-6b43-4492-a9ba-7b7eee992120-kube-api-access-m975m\") pod \"openshift-controller-manager-operator-756b6f6bc6-sm94r\" (UID: \"9fb614df-6b43-4492-a9ba-7b7eee992120\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.921226 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/beac9023-3506-4a56-9ac8-667172e2760d-srv-cert\") pod \"olm-operator-6b444d44fb-v94fb\" (UID: \"beac9023-3506-4a56-9ac8-667172e2760d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.921299 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/af3b37d7-7663-494c-a57a-419db7475661-node-bootstrap-token\") pod \"machine-config-server-rwd59\" (UID: \"af3b37d7-7663-494c-a57a-419db7475661\") " pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.921357 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-config-volume\") pod \"dns-default-n5p54\" (UID: \"6dcbb152-ea3e-4db5-aa09-d60e3768a99c\") " pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.921436 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/af3b37d7-7663-494c-a57a-419db7475661-certs\") pod \"machine-config-server-rwd59\" (UID: \"af3b37d7-7663-494c-a57a-419db7475661\") " pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.921491 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-metrics-tls\") pod \"dns-default-n5p54\" (UID: \"6dcbb152-ea3e-4db5-aa09-d60e3768a99c\") " pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.921564 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.921644 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/92906dea-785c-45c4-95fc-9e6ed15fc609-cert\") pod \"ingress-canary-rzxln\" (UID: \"92906dea-785c-45c4-95fc-9e6ed15fc609\") " pod="openshift-ingress-canary/ingress-canary-rzxln" Dec 11 08:39:18 crc kubenswrapper[4629]: W1211 08:39:18.922589 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-29413b786c22f310066a0a697144c21bdfa64d2a2814a650af0e1a653611992c WatchSource:0}: Error finding container 29413b786c22f310066a0a697144c21bdfa64d2a2814a650af0e1a653611992c: Status 404 returned error can't find the container with id 29413b786c22f310066a0a697144c21bdfa64d2a2814a650af0e1a653611992c Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.922648 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.918771 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"2ac0d8ec8fe4fc63e449faa2a226f6c4d4468fe257ecad9277dcce8b0bc4f642"} Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.923120 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"64b74c615bc6744d1b80bf778b5d291d73bf1aa81282e348b64b42f60613d1d7"} Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.923308 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bktzv"] Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.923410 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:39:18 crc kubenswrapper[4629]: E1211 08:39:18.923684 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:19.423670116 +0000 UTC m=+147.489085734 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.923714 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-config-volume\") pod \"dns-default-n5p54\" (UID: \"6dcbb152-ea3e-4db5-aa09-d60e3768a99c\") " pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.928466 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/92906dea-785c-45c4-95fc-9e6ed15fc609-cert\") pod \"ingress-canary-rzxln\" (UID: \"92906dea-785c-45c4-95fc-9e6ed15fc609\") " pod="openshift-ingress-canary/ingress-canary-rzxln" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.930422 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/af3b37d7-7663-494c-a57a-419db7475661-node-bootstrap-token\") pod \"machine-config-server-rwd59\" (UID: \"af3b37d7-7663-494c-a57a-419db7475661\") " pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.931946 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-metrics-tls\") pod \"dns-default-n5p54\" (UID: \"6dcbb152-ea3e-4db5-aa09-d60e3768a99c\") " pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.932282 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85h6k\" (UniqueName: \"kubernetes.io/projected/449eb3bb-7054-452d-85d6-558aa5efa682-kube-api-access-85h6k\") pod \"oauth-openshift-558db77b4-4jtwf\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.934671 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/af3b37d7-7663-494c-a57a-419db7475661-certs\") pod \"machine-config-server-rwd59\" (UID: \"af3b37d7-7663-494c-a57a-419db7475661\") " pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.942274 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.948480 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.949523 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/beac9023-3506-4a56-9ac8-667172e2760d-srv-cert\") pod \"olm-operator-6b444d44fb-v94fb\" (UID: \"beac9023-3506-4a56-9ac8-667172e2760d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.965803 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzqmt\" (UniqueName: \"kubernetes.io/projected/cca13586-fe9e-43e9-acb8-c3ff573ff1db-kube-api-access-tzqmt\") pod \"machine-approver-56656f9798-rrmqz\" (UID: \"cca13586-fe9e-43e9-acb8-c3ff573ff1db\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.980494 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2dzg\" (UniqueName: \"kubernetes.io/projected/c34dd798-968c-4df8-bd03-8308d64fcd8c-kube-api-access-f2dzg\") pod \"apiserver-7bbb656c7d-m5fhn\" (UID: \"c34dd798-968c-4df8-bd03-8308d64fcd8c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.994036 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.994365 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv" Dec 11 08:39:18 crc kubenswrapper[4629]: I1211 08:39:18.999710 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz7w8\" (UniqueName: \"kubernetes.io/projected/9311bda6-d765-4970-9749-3fc2b826b0a3-kube-api-access-hz7w8\") pod \"dns-operator-744455d44c-5ccrx\" (UID: \"9311bda6-d765-4970-9749-3fc2b826b0a3\") " pod="openshift-dns-operator/dns-operator-744455d44c-5ccrx" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.004459 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.023186 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.026325 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:19 crc kubenswrapper[4629]: E1211 08:39:19.030102 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:19.530080807 +0000 UTC m=+147.595496425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.033834 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd"] Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.034164 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2lmf\" (UniqueName: \"kubernetes.io/projected/a33a0903-3d69-4017-b274-4f188c6b89c8-kube-api-access-b2lmf\") pod \"ingress-operator-5b745b69d9-pvsn8\" (UID: \"a33a0903-3d69-4017-b274-4f188c6b89c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.044071 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbrrd\" (UniqueName: \"kubernetes.io/projected/ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb-kube-api-access-nbrrd\") pod \"packageserver-d55dfcdfc-7gkf5\" (UID: \"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.044437 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-5ccrx" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.053145 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg4mm\" (UniqueName: \"kubernetes.io/projected/ec8356f9-5138-4fe5-867d-22d4d636c0d9-kube-api-access-gg4mm\") pod \"etcd-operator-b45778765-nn9r8\" (UID: \"ec8356f9-5138-4fe5-867d-22d4d636c0d9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.065608 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx4fp\" (UniqueName: \"kubernetes.io/projected/aa821340-be6f-4719-8379-c4e6eeda38aa-kube-api-access-nx4fp\") pod \"service-ca-operator-777779d784-h5fwd\" (UID: \"aa821340-be6f-4719-8379-c4e6eeda38aa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.068545 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" Dec 11 08:39:19 crc kubenswrapper[4629]: W1211 08:39:19.074690 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9167ddcb_69c2_4861_80e2_1ab4419b8e93.slice/crio-023dc4f18d7451585dc5dd3e0465391614ed3838f75738f175cd33a99da3585d WatchSource:0}: Error finding container 023dc4f18d7451585dc5dd3e0465391614ed3838f75738f175cd33a99da3585d: Status 404 returned error can't find the container with id 023dc4f18d7451585dc5dd3e0465391614ed3838f75738f175cd33a99da3585d Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.100791 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc96z\" (UniqueName: \"kubernetes.io/projected/695005e9-0af9-444d-8117-823af0b1ed46-kube-api-access-lc96z\") pod \"machine-config-controller-84d6567774-mbgh9\" (UID: \"695005e9-0af9-444d-8117-823af0b1ed46\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.116214 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh7kz\" (UniqueName: \"kubernetes.io/projected/641fc441-d314-4b07-9fdf-98a81d21744f-kube-api-access-dh7kz\") pod \"router-default-5444994796-r9x7m\" (UID: \"641fc441-d314-4b07-9fdf-98a81d21744f\") " pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.120794 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.127657 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.128710 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:19 crc kubenswrapper[4629]: E1211 08:39:19.131557 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:19.631539259 +0000 UTC m=+147.696954867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.141536 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a33a0903-3d69-4017-b274-4f188c6b89c8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pvsn8\" (UID: \"a33a0903-3d69-4017-b274-4f188c6b89c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.164973 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2c7tp"] Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.172657 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkkmx\" (UniqueName: \"kubernetes.io/projected/83f7a7a2-12d8-43b0-9a3a-3926d3f0092e-kube-api-access-vkkmx\") pod \"downloads-7954f5f757-wbrg7\" (UID: \"83f7a7a2-12d8-43b0-9a3a-3926d3f0092e\") " pod="openshift-console/downloads-7954f5f757-wbrg7" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.176367 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttxmt\" (UniqueName: \"kubernetes.io/projected/beac9023-3506-4a56-9ac8-667172e2760d-kube-api-access-ttxmt\") pod \"olm-operator-6b444d44fb-v94fb\" (UID: \"beac9023-3506-4a56-9ac8-667172e2760d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.215574 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.216395 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.217282 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.237767 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:19 crc kubenswrapper[4629]: E1211 08:39:19.238118 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:19.738104424 +0000 UTC m=+147.803520032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.239579 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6v2n\" (UniqueName: \"kubernetes.io/projected/d44b0b79-360a-4d50-abb1-ef1ded04c565-kube-api-access-w6v2n\") pod \"csi-hostpathplugin-wdcs7\" (UID: \"d44b0b79-360a-4d50-abb1-ef1ded04c565\") " pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.240800 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-lhszp"] Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.242809 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gggk9\" (UniqueName: \"kubernetes.io/projected/0463664b-27fa-42ed-8369-442d93fbdcb5-kube-api-access-gggk9\") pod \"package-server-manager-789f6589d5-8rlll\" (UID: \"0463664b-27fa-42ed-8369-442d93fbdcb5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.244137 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db4ss\" (UniqueName: \"kubernetes.io/projected/aedfd32e-f597-4753-8fbc-824b562b9c77-kube-api-access-db4ss\") pod \"control-plane-machine-set-operator-78cbb6b69f-m9222\" (UID: \"aedfd32e-f597-4753-8fbc-824b562b9c77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222" Dec 11 08:39:19 crc kubenswrapper[4629]: W1211 08:39:19.244943 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d819e7a_6f7a_4f6e_bbba_4b334546f172.slice/crio-ea5b81c4e7c45ef5d714258359fd60e70ca8d5febeb8e08f62cff7dfd65067ec WatchSource:0}: Error finding container ea5b81c4e7c45ef5d714258359fd60e70ca8d5febeb8e08f62cff7dfd65067ec: Status 404 returned error can't find the container with id ea5b81c4e7c45ef5d714258359fd60e70ca8d5febeb8e08f62cff7dfd65067ec Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.247154 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z6bv\" (UniqueName: \"kubernetes.io/projected/3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8-kube-api-access-4z6bv\") pod \"multus-admission-controller-857f4d67dd-fmmgm\" (UID: \"3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fmmgm" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.248409 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fmmgm" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.281624 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.285045 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcdst\" (UniqueName: \"kubernetes.io/projected/9e92bcf9-0313-43ca-b2ae-fbdd9586aea2-kube-api-access-dcdst\") pod \"kube-storage-version-migrator-operator-b67b599dd-rw8lc\" (UID: \"9e92bcf9-0313-43ca-b2ae-fbdd9586aea2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.288336 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.303770 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.303774 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twtj4\" (UniqueName: \"kubernetes.io/projected/07febd9a-7813-4722-af4d-9b165be26662-kube-api-access-twtj4\") pod \"service-ca-9c57cc56f-6qgqv\" (UID: \"07febd9a-7813-4722-af4d-9b165be26662\") " pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.313446 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlk28\" (UniqueName: \"kubernetes.io/projected/8bf66625-0e08-4c99-9633-85558c1ce12f-kube-api-access-zlk28\") pod \"machine-config-operator-74547568cd-g5p2r\" (UID: \"8bf66625-0e08-4c99-9633-85558c1ce12f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.322766 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-wbrg7" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.331124 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhbv9\" (UniqueName: \"kubernetes.io/projected/bfbd1947-e471-47dd-8c36-916c28cfaf79-kube-api-access-dhbv9\") pod \"migrator-59844c95c7-mbkcz\" (UID: \"bfbd1947-e471-47dd-8c36-916c28cfaf79\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mbkcz" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.345661 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:19 crc kubenswrapper[4629]: E1211 08:39:19.346031 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:19.846019202 +0000 UTC m=+147.911434810 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.348914 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.353006 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74rkg\" (UniqueName: \"kubernetes.io/projected/af3b37d7-7663-494c-a57a-419db7475661-kube-api-access-74rkg\") pod \"machine-config-server-rwd59\" (UID: \"af3b37d7-7663-494c-a57a-419db7475661\") " pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.356636 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t7tg7"] Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.388412 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.392124 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e84237c3-8fcc-4a25-9d54-2422dcc628c5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8v64\" (UID: \"e84237c3-8fcc-4a25-9d54-2422dcc628c5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.399083 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mbkcz" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.399493 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzqql\" (UniqueName: \"kubernetes.io/projected/6dcbb152-ea3e-4db5-aa09-d60e3768a99c-kube-api-access-bzqql\") pod \"dns-default-n5p54\" (UID: \"6dcbb152-ea3e-4db5-aa09-d60e3768a99c\") " pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.407029 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r"] Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.408303 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.412559 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lr8q8\" (UID: \"98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.422276 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.427230 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d86l\" (UniqueName: \"kubernetes.io/projected/92906dea-785c-45c4-95fc-9e6ed15fc609-kube-api-access-7d86l\") pod \"ingress-canary-rzxln\" (UID: \"92906dea-785c-45c4-95fc-9e6ed15fc609\") " pod="openshift-ingress-canary/ingress-canary-rzxln" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.439556 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.448331 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:19 crc kubenswrapper[4629]: E1211 08:39:19.448734 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:19.948719755 +0000 UTC m=+148.014135363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.449142 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.459647 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkqmk\" (UniqueName: \"kubernetes.io/projected/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-kube-api-access-tkqmk\") pod \"collect-profiles-29424030-vtnxm\" (UID: \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.483122 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfqpc\" (UniqueName: \"kubernetes.io/projected/3efc8f47-f897-4072-b1ea-da866707146c-kube-api-access-nfqpc\") pod \"catalog-operator-68c6474976-x9fpk\" (UID: \"3efc8f47-f897-4072-b1ea-da866707146c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.485743 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.491798 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv9cf\" (UniqueName: \"kubernetes.io/projected/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-kube-api-access-xv9cf\") pod \"marketplace-operator-79b997595-rzqjs\" (UID: \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\") " pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.496523 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.516565 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2d4978b-296a-4177-955e-d367860c19ab-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-947kd\" (UID: \"d2d4978b-296a-4177-955e-d367860c19ab\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.524651 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.533651 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.559975 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:19 crc kubenswrapper[4629]: E1211 08:39:19.560313 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:20.06029949 +0000 UTC m=+148.125715098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.566041 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.576740 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rzxln" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.612634 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rwd59" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.620223 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.650372 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4jtwf"] Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.665051 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.665191 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:19 crc kubenswrapper[4629]: E1211 08:39:19.665577 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:20.165557563 +0000 UTC m=+148.230973171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.678816 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.770500 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:19 crc kubenswrapper[4629]: E1211 08:39:19.771041 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:20.271026824 +0000 UTC m=+148.336442512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:19 crc kubenswrapper[4629]: W1211 08:39:19.833398 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod449eb3bb_7054_452d_85d6_558aa5efa682.slice/crio-4da6ae0bfa62287ccc5e9a0da29741bb282b3cb062a462e63ed1e58610ef2f6d WatchSource:0}: Error finding container 4da6ae0bfa62287ccc5e9a0da29741bb282b3cb062a462e63ed1e58610ef2f6d: Status 404 returned error can't find the container with id 4da6ae0bfa62287ccc5e9a0da29741bb282b3cb062a462e63ed1e58610ef2f6d Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.872391 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:19 crc kubenswrapper[4629]: E1211 08:39:19.872705 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:20.372690071 +0000 UTC m=+148.438105679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.923194 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv"] Dec 11 08:39:19 crc kubenswrapper[4629]: I1211 08:39:19.981163 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:19 crc kubenswrapper[4629]: E1211 08:39:19.981543 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:20.481525139 +0000 UTC m=+148.546940747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.126958 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:20 crc kubenswrapper[4629]: E1211 08:39:20.127132 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:20.627092361 +0000 UTC m=+148.692507969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.127510 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:20 crc kubenswrapper[4629]: E1211 08:39:20.127920 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:20.627907127 +0000 UTC m=+148.693322735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.155920 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" event={"ID":"dbb99da1-1049-4e03-bc19-49988fc37947","Type":"ContainerStarted","Data":"f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.155967 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" event={"ID":"dbb99da1-1049-4e03-bc19-49988fc37947","Type":"ContainerStarted","Data":"fcdeddb96f242d5fec813d5cecdf37a59ac447ba693c97d91492a0e9c2a40e14"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.155986 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.197037 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" event={"ID":"449eb3bb-7054-452d-85d6-558aa5efa682","Type":"ContainerStarted","Data":"4da6ae0bfa62287ccc5e9a0da29741bb282b3cb062a462e63ed1e58610ef2f6d"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.202387 4629 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-2c7tp container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.202443 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" podUID="dbb99da1-1049-4e03-bc19-49988fc37947" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.223524 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" event={"ID":"cca13586-fe9e-43e9-acb8-c3ff573ff1db","Type":"ContainerStarted","Data":"c39edc8a96467b9ede57890832917a97ef41601fe64316c1b5c11034843bb7ba"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.224777 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nn9r8"] Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.224925 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8q9kf"] Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.231601 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:20 crc kubenswrapper[4629]: E1211 08:39:20.233226 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:20.733203452 +0000 UTC m=+148.798619060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.262195 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" event={"ID":"9167ddcb-69c2-4861-80e2-1ab4419b8e93","Type":"ContainerStarted","Data":"e9e4c5418b3b30d7580d9dedae5588e4481d9c3a67bd795dc325a494217a2d40"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.262243 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" event={"ID":"9167ddcb-69c2-4861-80e2-1ab4419b8e93","Type":"ContainerStarted","Data":"023dc4f18d7451585dc5dd3e0465391614ed3838f75738f175cd33a99da3585d"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.281485 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"79a481a800cc4096a573cda49dfefa5062feb8b4fea86d6cb3318293eafba66c"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.281548 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"29413b786c22f310066a0a697144c21bdfa64d2a2814a650af0e1a653611992c"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.283614 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-npk4p"] Dec 11 08:39:20 crc kubenswrapper[4629]: W1211 08:39:20.293430 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf3b37d7_7663_494c_a57a_419db7475661.slice/crio-532b81725d7649da9b615276a7bac2d596b3af309cf079ee773de63d204534df WatchSource:0}: Error finding container 532b81725d7649da9b615276a7bac2d596b3af309cf079ee773de63d204534df: Status 404 returned error can't find the container with id 532b81725d7649da9b615276a7bac2d596b3af309cf079ee773de63d204534df Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.299131 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" event={"ID":"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a","Type":"ContainerStarted","Data":"f551b49825d84ed6fda5cab18fbd352505e30cd3be62e5c07f44b3948b493486"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.299183 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" event={"ID":"fdb51e07-8d01-4c36-9e03-68c0dbc82c0a","Type":"ContainerStarted","Data":"ed737f531089e33933ce7b08cb6304d48bedbec794e0de869ea2c10496baaee6"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.326959 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5"] Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.337243 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:20 crc kubenswrapper[4629]: E1211 08:39:20.337708 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:20.83769135 +0000 UTC m=+148.903106958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.352801 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r9x7m" event={"ID":"641fc441-d314-4b07-9fdf-98a81d21744f","Type":"ContainerStarted","Data":"13ea846678056374a0a8c1fb9e2600f560a769ae15d1b1e1a0bb3ba448ebd70a"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.360287 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-lhszp" event={"ID":"9d819e7a-6f7a-4f6e-bbba-4b334546f172","Type":"ContainerStarted","Data":"751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.360342 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-lhszp" event={"ID":"9d819e7a-6f7a-4f6e-bbba-4b334546f172","Type":"ContainerStarted","Data":"ea5b81c4e7c45ef5d714258359fd60e70ca8d5febeb8e08f62cff7dfd65067ec"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.374876 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" event={"ID":"9fb614df-6b43-4492-a9ba-7b7eee992120","Type":"ContainerStarted","Data":"c81fc09a8e6eac3b62c381ff8dbc80576d7a78caf189d9c8dc59ba5606ebb283"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.378958 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" event={"ID":"0c0faec1-d9b4-49c4-8595-381e5d0c4a56","Type":"ContainerStarted","Data":"11d78f4f6f9130fbba7e0acadd64209173976ae80cca3d34108682911f93ea83"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.381692 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"55a2cf0427d3b1325a5a856a768cbd695731b1698c773b6ea59b7cd28fb3edca"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.381751 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"da4e301919d822b6bf31bb36969c0aab4fd36393816a458785ea4fd7c100c6ab"} Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.403685 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v"] Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.438744 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:20 crc kubenswrapper[4629]: E1211 08:39:20.439279 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:20.939259636 +0000 UTC m=+149.004675244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.440663 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:20 crc kubenswrapper[4629]: E1211 08:39:20.444668 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:20.944652988 +0000 UTC m=+149.010068596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.488313 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.488388 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.546675 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:20 crc kubenswrapper[4629]: E1211 08:39:20.551745 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:21.05172629 +0000 UTC m=+149.117141898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.651658 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:20 crc kubenswrapper[4629]: E1211 08:39:20.652340 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:21.152328855 +0000 UTC m=+149.217744463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.709519 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd"] Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.710773 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5"] Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.721836 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t7dc6" podStartSLOduration=129.721802874 podStartE2EDuration="2m9.721802874s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:20.709353006 +0000 UTC m=+148.774768614" watchObservedRunningTime="2025-12-11 08:39:20.721802874 +0000 UTC m=+148.787218472" Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.763416 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:20 crc kubenswrapper[4629]: E1211 08:39:20.763708 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:21.263690613 +0000 UTC m=+149.329106221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.775858 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fmmgm"] Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.804134 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9"] Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.824047 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5ccrx"] Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.866518 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:20 crc kubenswrapper[4629]: E1211 08:39:20.866975 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:21.366962653 +0000 UTC m=+149.432378261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.967098 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:20 crc kubenswrapper[4629]: E1211 08:39:20.967469 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:21.467438083 +0000 UTC m=+149.532853691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:20 crc kubenswrapper[4629]: I1211 08:39:20.967637 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:20 crc kubenswrapper[4629]: E1211 08:39:20.968018 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:21.468009802 +0000 UTC m=+149.533425470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.068285 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:21 crc kubenswrapper[4629]: E1211 08:39:21.068683 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:21.568667488 +0000 UTC m=+149.634083096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.169530 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:21 crc kubenswrapper[4629]: E1211 08:39:21.170147 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:21.6701308 +0000 UTC m=+149.735546408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.270717 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:21 crc kubenswrapper[4629]: E1211 08:39:21.271189 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:21.771175319 +0000 UTC m=+149.836590927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.375279 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:21 crc kubenswrapper[4629]: E1211 08:39:21.375639 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:21.875625116 +0000 UTC m=+149.941040724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.386915 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" podStartSLOduration=130.386893566 podStartE2EDuration="2m10.386893566s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:21.33565838 +0000 UTC m=+149.401073998" watchObservedRunningTime="2025-12-11 08:39:21.386893566 +0000 UTC m=+149.452309174" Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.476326 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:21 crc kubenswrapper[4629]: E1211 08:39:21.476691 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:21.976674425 +0000 UTC m=+150.042090033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.509783 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" event={"ID":"ec0bf78c-4e35-4129-8976-c8d805991843","Type":"ContainerStarted","Data":"0052ff9cceb9ed9b182481d4b7d1dc60eb7b6265d2bed65849e1826f40b7ee0f"} Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.545454 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8q9kf" event={"ID":"d60662db-4515-4c66-8bfd-e418e6f258fa","Type":"ContainerStarted","Data":"c80e6373db5c2c5efab62d09c90ba3c9a41c545127a3b6a4f6a74c68dec14007"} Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.545506 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8q9kf" event={"ID":"d60662db-4515-4c66-8bfd-e418e6f258fa","Type":"ContainerStarted","Data":"ade282b11fa1011c6bc38e51ac013922be86d6e96dc2e5e09d13630bb403d09c"} Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.546401 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.578638 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:21 crc kubenswrapper[4629]: E1211 08:39:21.579177 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:22.07915922 +0000 UTC m=+150.144574828 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.596700 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" event={"ID":"ec8356f9-5138-4fe5-867d-22d4d636c0d9","Type":"ContainerStarted","Data":"09193f89975ae0e453c87bfd46f1c0ef66c8ed54c2495159ebd6cd27bbe7bcf5"} Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.614237 4629 patch_prober.go:28] interesting pod/console-operator-58897d9998-8q9kf container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.614288 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-8q9kf" podUID="d60662db-4515-4c66-8bfd-e418e6f258fa" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.616004 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-bktzv" podStartSLOduration=130.615992168 podStartE2EDuration="2m10.615992168s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:21.615105119 +0000 UTC m=+149.680520727" watchObservedRunningTime="2025-12-11 08:39:21.615992168 +0000 UTC m=+149.681407776" Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.617234 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-lhszp" podStartSLOduration=130.617227887 podStartE2EDuration="2m10.617227887s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:21.558104797 +0000 UTC m=+149.623520415" watchObservedRunningTime="2025-12-11 08:39:21.617227887 +0000 UTC m=+149.682643495" Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.681496 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:21 crc kubenswrapper[4629]: E1211 08:39:21.682604 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:22.182559244 +0000 UTC m=+150.247974852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.698191 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r9x7m" event={"ID":"641fc441-d314-4b07-9fdf-98a81d21744f","Type":"ContainerStarted","Data":"b958cfb37e4a0498c892dc8eab852487b74ab0da2de67e58560e1a96b8357a12"} Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.773667 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" event={"ID":"aa821340-be6f-4719-8379-c4e6eeda38aa","Type":"ContainerStarted","Data":"89917a0f0351acc19332f02596777d6d2e027ad43970da8896284bafed87ed67"} Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.800536 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bnswd" podStartSLOduration=130.800520914 podStartE2EDuration="2m10.800520914s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:21.77038937 +0000 UTC m=+149.835804978" watchObservedRunningTime="2025-12-11 08:39:21.800520914 +0000 UTC m=+149.865936522" Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.800646 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:21 crc kubenswrapper[4629]: E1211 08:39:21.800969 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:22.300957168 +0000 UTC m=+150.366372776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.804840 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wdcs7"] Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.838285 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" event={"ID":"cca13586-fe9e-43e9-acb8-c3ff573ff1db","Type":"ContainerStarted","Data":"60060de4a6e9b2ef3c48a42fe5b187a6fd55c3fcb1be2696be7073ca3ed6bfd2"} Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.867637 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" event={"ID":"9fb614df-6b43-4492-a9ba-7b7eee992120","Type":"ContainerStarted","Data":"c140b2b10d8e9889a0a14dd799b906592e50798f5bbd52ca2f0a229a0907dc35"} Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.936562 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:21 crc kubenswrapper[4629]: E1211 08:39:21.938385 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:22.438362528 +0000 UTC m=+150.503778136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.942761 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rwd59" event={"ID":"af3b37d7-7663-494c-a57a-419db7475661","Type":"ContainerStarted","Data":"7fc824fe1cadc6576bb9a55026c8833cb680d16acec6507b4881fcb6caaa2a41"} Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.942803 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rwd59" event={"ID":"af3b37d7-7663-494c-a57a-419db7475661","Type":"ContainerStarted","Data":"532b81725d7649da9b615276a7bac2d596b3af309cf079ee773de63d204534df"} Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.978672 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv" event={"ID":"28a9a7b5-a9ca-4ac4-bef4-da3715a27b67","Type":"ContainerStarted","Data":"d48818324aa5172686fdc222c9643af6ec77d2a39d9b09dda8972b37d3a4a7cc"} Dec 11 08:39:21 crc kubenswrapper[4629]: I1211 08:39:21.983679 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-r9x7m" podStartSLOduration=130.983653565 podStartE2EDuration="2m10.983653565s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:21.904680281 +0000 UTC m=+149.970095889" watchObservedRunningTime="2025-12-11 08:39:21.983653565 +0000 UTC m=+150.049069173" Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.014947 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-wbrg7"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.038945 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:22 crc kubenswrapper[4629]: E1211 08:39:22.039275 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:22.539263492 +0000 UTC m=+150.604679100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.056051 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" event={"ID":"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb","Type":"ContainerStarted","Data":"5d482ee240e882818d5fb19db75f5091fb481ed98ebdbd03c30e4fa48cd58b52"} Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.063637 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" event={"ID":"ed370895-f652-499d-8382-69fb05f78f03","Type":"ContainerStarted","Data":"0ee4c860c5f557b8836390bfa615718f4c900dc5c113ea0c5e39db15630b6487"} Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.131093 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" event={"ID":"d957c3d0-bff4-4c92-82ea-3900ea7efa06","Type":"ContainerStarted","Data":"ab67bd2ad21a45c5c945b9f3878d2b39d1d190bd441a8fdc022cf245b5d3e675"} Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.139434 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:22 crc kubenswrapper[4629]: E1211 08:39:22.140593 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:22.64057477 +0000 UTC m=+150.705990388 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.155467 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6qgqv"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.160031 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" event={"ID":"695005e9-0af9-444d-8117-823af0b1ed46","Type":"ContainerStarted","Data":"6abf2a1471cfdb2636d8501cfd4c823df1c76ec606a382d529e16f395d64ba7d"} Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.161620 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-8q9kf" podStartSLOduration=131.161601642 podStartE2EDuration="2m11.161601642s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:22.127986058 +0000 UTC m=+150.193401666" watchObservedRunningTime="2025-12-11 08:39:22.161601642 +0000 UTC m=+150.227017250" Dec 11 08:39:22 crc kubenswrapper[4629]: W1211 08:39:22.178407 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07febd9a_7813_4722_af4d_9b165be26662.slice/crio-6b415c7fb7b4ee774a02393118a736145c6d8a060606d035ce2fd3676f54b7a5 WatchSource:0}: Error finding container 6b415c7fb7b4ee774a02393118a736145c6d8a060606d035ce2fd3676f54b7a5: Status 404 returned error can't find the container with id 6b415c7fb7b4ee774a02393118a736145c6d8a060606d035ce2fd3676f54b7a5 Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.193250 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fmmgm" event={"ID":"3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8","Type":"ContainerStarted","Data":"c16bedcc7ac540a08961e6afe0179a146a93fb271b3a947bc2e1a3ae83347d3f"} Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.249556 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:22 crc kubenswrapper[4629]: E1211 08:39:22.250040 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:22.750028417 +0000 UTC m=+150.815444025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.253163 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5ccrx" event={"ID":"9311bda6-d765-4970-9749-3fc2b826b0a3","Type":"ContainerStarted","Data":"17f86b0cacc5f06c90a54ca870e66b5c42d13399f1c7d301f1ef27c4a7ec0b27"} Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.258752 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.258790 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.258803 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.258813 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.284761 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.297530 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rzqjs"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.357301 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.359336 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:22 crc kubenswrapper[4629]: E1211 08:39:22.364515 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:22.864491565 +0000 UTC m=+150.929907343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.379143 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-n5p54"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.379955 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-rwd59" podStartSLOduration=6.379930588 podStartE2EDuration="6.379930588s" podCreationTimestamp="2025-12-11 08:39:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:22.346758669 +0000 UTC m=+150.412174277" watchObservedRunningTime="2025-12-11 08:39:22.379930588 +0000 UTC m=+150.445346196" Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.407268 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:22 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:22 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:22 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.407336 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:22 crc kubenswrapper[4629]: W1211 08:39:22.413222 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc34dd798_968c_4df8_bd03_8308d64fcd8c.slice/crio-4d12937ed1002326f5362b72adc40eab55921f6060fbcc52ca36dd30dfbe51d7 WatchSource:0}: Error finding container 4d12937ed1002326f5362b72adc40eab55921f6060fbcc52ca36dd30dfbe51d7: Status 404 returned error can't find the container with id 4d12937ed1002326f5362b72adc40eab55921f6060fbcc52ca36dd30dfbe51d7 Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.445885 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.458953 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.462753 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:22 crc kubenswrapper[4629]: E1211 08:39:22.463236 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:22.96321803 +0000 UTC m=+151.028633638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.490032 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sm94r" podStartSLOduration=131.489984384 podStartE2EDuration="2m11.489984384s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:22.435814864 +0000 UTC m=+150.501230472" watchObservedRunningTime="2025-12-11 08:39:22.489984384 +0000 UTC m=+150.555399992" Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.493265 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.504020 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.521072 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.566814 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:22 crc kubenswrapper[4629]: E1211 08:39:22.567424 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:23.067409649 +0000 UTC m=+151.132825257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.611478 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.689699 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:22 crc kubenswrapper[4629]: E1211 08:39:22.690131 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:23.19011876 +0000 UTC m=+151.255534368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.704838 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd"] Dec 11 08:39:22 crc kubenswrapper[4629]: W1211 08:39:22.723589 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99df09c3_eae2_4db8_a82e_a7e1d1c462e5.slice/crio-4dc746562a886a2e349437170dc4d9a11e6e1bc50ae8b6642429882436139e44 WatchSource:0}: Error finding container 4dc746562a886a2e349437170dc4d9a11e6e1bc50ae8b6642429882436139e44: Status 404 returned error can't find the container with id 4dc746562a886a2e349437170dc4d9a11e6e1bc50ae8b6642429882436139e44 Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.753440 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-mbkcz"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.756895 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64"] Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.793236 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:22 crc kubenswrapper[4629]: E1211 08:39:22.793516 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:23.293496133 +0000 UTC m=+151.358911741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.817742 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rzxln"] Dec 11 08:39:22 crc kubenswrapper[4629]: W1211 08:39:22.858012 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode84237c3_8fcc_4a25_9d54_2422dcc628c5.slice/crio-35a9408e93ad64356c58570fa9de8669469e300064772f2b2e68e6c983e705db WatchSource:0}: Error finding container 35a9408e93ad64356c58570fa9de8669469e300064772f2b2e68e6c983e705db: Status 404 returned error can't find the container with id 35a9408e93ad64356c58570fa9de8669469e300064772f2b2e68e6c983e705db Dec 11 08:39:22 crc kubenswrapper[4629]: I1211 08:39:22.900878 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:22 crc kubenswrapper[4629]: E1211 08:39:22.901137 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:23.401127263 +0000 UTC m=+151.466542871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.001875 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:23 crc kubenswrapper[4629]: E1211 08:39:23.002295 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:23.502279264 +0000 UTC m=+151.567694872 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.104759 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:23 crc kubenswrapper[4629]: E1211 08:39:23.105173 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:23.605152582 +0000 UTC m=+151.670568190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.208156 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:23 crc kubenswrapper[4629]: E1211 08:39:23.208887 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:23.708869246 +0000 UTC m=+151.774284854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.326578 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:23 crc kubenswrapper[4629]: E1211 08:39:23.327065 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:23.827050062 +0000 UTC m=+151.892465660 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.361048 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:23 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:23 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:23 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.361102 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.410365 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" event={"ID":"0463664b-27fa-42ed-8369-442d93fbdcb5","Type":"ContainerStarted","Data":"3256af5e8884999a35fb1e0729f28cff294c7c562fd7e9f0375b698d916d9f72"} Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.438030 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:23 crc kubenswrapper[4629]: E1211 08:39:23.438288 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:23.938270767 +0000 UTC m=+152.003686375 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.439186 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" event={"ID":"c34dd798-968c-4df8-bd03-8308d64fcd8c","Type":"ContainerStarted","Data":"4d12937ed1002326f5362b72adc40eab55921f6060fbcc52ca36dd30dfbe51d7"} Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.494907 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" event={"ID":"ec8356f9-5138-4fe5-867d-22d4d636c0d9","Type":"ContainerStarted","Data":"286c9018a0f448f709518ecd32c0481f2414acde39b49197a525f65424936b8a"} Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.513182 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" event={"ID":"449eb3bb-7054-452d-85d6-558aa5efa682","Type":"ContainerStarted","Data":"94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059"} Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.513932 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.535176 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" event={"ID":"3efc8f47-f897-4072-b1ea-da866707146c","Type":"ContainerStarted","Data":"d051356d0e60812cb0dc81d32715f67510cf09aaf84a4840f849932403e366a7"} Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.542667 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:23 crc kubenswrapper[4629]: E1211 08:39:23.543132 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:24.043116966 +0000 UTC m=+152.108532574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.562134 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" event={"ID":"695005e9-0af9-444d-8117-823af0b1ed46","Type":"ContainerStarted","Data":"18c7e2273ee2784078326717129453e7217ecbb3b0876c02726ffd80abe74636"} Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.562189 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" event={"ID":"695005e9-0af9-444d-8117-823af0b1ed46","Type":"ContainerStarted","Data":"c14355aed87b61e42638b20068cb149766aa77d89c7c2c8e9da14f813389d525"} Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.573254 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fmmgm" event={"ID":"3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8","Type":"ContainerStarted","Data":"e8c9ec6399fb7e1ee58dffba29c980112fa1b86d6434639b17d06f56907fe175"} Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.649328 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:23 crc kubenswrapper[4629]: E1211 08:39:23.651146 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:24.151120598 +0000 UTC m=+152.216536206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.767933 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" event={"ID":"beac9023-3506-4a56-9ac8-667172e2760d","Type":"ContainerStarted","Data":"0a7deff0f5e32a7d9647858e30f502a3d8ea6273daa4653670925e3c98d811e2"} Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.774039 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:23 crc kubenswrapper[4629]: E1211 08:39:23.775580 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:24.275563034 +0000 UTC m=+152.340978642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.800314 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" podStartSLOduration=132.800295864 podStartE2EDuration="2m12.800295864s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:23.799464158 +0000 UTC m=+151.864879766" watchObservedRunningTime="2025-12-11 08:39:23.800295864 +0000 UTC m=+151.865711472" Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.845430 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" event={"ID":"ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb","Type":"ContainerStarted","Data":"3b2a5b7880149a1ce1a69c27b1062100bdf28a87471a42339355443f7196f0e0"} Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.846562 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.869391 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mbgh9" podStartSLOduration=132.869373732 podStartE2EDuration="2m12.869373732s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:23.866831211 +0000 UTC m=+151.932246829" watchObservedRunningTime="2025-12-11 08:39:23.869373732 +0000 UTC m=+151.934789340" Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.876752 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.878120 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" event={"ID":"aa821340-be6f-4719-8379-c4e6eeda38aa","Type":"ContainerStarted","Data":"f0cb80be6017a3b53aacae70a914e42d521d6cbbbf13f4fc121497544398abba"} Dec 11 08:39:23 crc kubenswrapper[4629]: E1211 08:39:23.878795 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:24.378772612 +0000 UTC m=+152.444188220 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.934222 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" event={"ID":"07febd9a-7813-4722-af4d-9b165be26662","Type":"ContainerStarted","Data":"6b415c7fb7b4ee774a02393118a736145c6d8a060606d035ce2fd3676f54b7a5"} Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.958077 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-nn9r8" podStartSLOduration=132.958060456 podStartE2EDuration="2m12.958060456s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:23.957539769 +0000 UTC m=+152.022955377" watchObservedRunningTime="2025-12-11 08:39:23.958060456 +0000 UTC m=+152.023476064" Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.983976 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:23 crc kubenswrapper[4629]: E1211 08:39:23.985880 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:24.485861364 +0000 UTC m=+152.551276982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:23 crc kubenswrapper[4629]: I1211 08:39:23.986756 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" event={"ID":"d44b0b79-360a-4d50-abb1-ef1ded04c565","Type":"ContainerStarted","Data":"c66132d23cebdbf5b70c74a3da00e4a80ef3643600ee0b3fab657aa3e1a192d7"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.013658 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" podStartSLOduration=132.013640082 podStartE2EDuration="2m12.013640082s" podCreationTimestamp="2025-12-11 08:37:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:24.011108391 +0000 UTC m=+152.076523999" watchObservedRunningTime="2025-12-11 08:39:24.013640082 +0000 UTC m=+152.079055690" Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.021683 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222" event={"ID":"aedfd32e-f597-4753-8fbc-824b562b9c77","Type":"ContainerStarted","Data":"adb4ef1aa51c48e2806c2cb454ecc842a796b19272ef0f5e846e73ec133e414e"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.060443 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" event={"ID":"98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55","Type":"ContainerStarted","Data":"efd207583ccfdbfe17dd7349a7274c29e98692d44c0a08f72fa05341d0c2dc74"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.106882 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:24 crc kubenswrapper[4629]: E1211 08:39:24.108234 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:24.608210953 +0000 UTC m=+152.673626561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.124557 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" event={"ID":"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc","Type":"ContainerStarted","Data":"39e0d635f9c4c3f1a4f983b4b537ef0a1f925ec3e32c1602829456e2f268ae56"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.165070 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" event={"ID":"99df09c3-eae2-4db8-a82e-a7e1d1c462e5","Type":"ContainerStarted","Data":"4dc746562a886a2e349437170dc4d9a11e6e1bc50ae8b6642429882436139e44"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.199983 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" event={"ID":"9e92bcf9-0313-43ca-b2ae-fbdd9586aea2","Type":"ContainerStarted","Data":"45cf51e5f5e0f25d0176e44ee41a604bcd49129ac6ad321e91a29a1433db0d93"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.209089 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:24 crc kubenswrapper[4629]: E1211 08:39:24.209780 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:24.709766479 +0000 UTC m=+152.775182087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.231980 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h5fwd" podStartSLOduration=132.231964008 podStartE2EDuration="2m12.231964008s" podCreationTimestamp="2025-12-11 08:37:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:24.112563143 +0000 UTC m=+152.177978751" watchObservedRunningTime="2025-12-11 08:39:24.231964008 +0000 UTC m=+152.297379616" Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.284970 4629 generic.go:334] "Generic (PLEG): container finished" podID="d957c3d0-bff4-4c92-82ea-3900ea7efa06" containerID="0f91c3d1540175f560d803176cb75ebdd1ef7775299ede09950113dd73b36dd3" exitCode=0 Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.307360 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" event={"ID":"d957c3d0-bff4-4c92-82ea-3900ea7efa06","Type":"ContainerDied","Data":"0f91c3d1540175f560d803176cb75ebdd1ef7775299ede09950113dd73b36dd3"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.311660 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:24 crc kubenswrapper[4629]: E1211 08:39:24.311992 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:24.811947824 +0000 UTC m=+152.877363432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.312229 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:24 crc kubenswrapper[4629]: E1211 08:39:24.312700 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:24.812692277 +0000 UTC m=+152.878107875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.332108 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-n5p54" event={"ID":"6dcbb152-ea3e-4db5-aa09-d60e3768a99c","Type":"ContainerStarted","Data":"4c0310f32289dd58f72e66efa7847336518bcecdfdc069325eec53d9987c4dee"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.355514 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:24 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:24 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:24 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.355588 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.356277 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" event={"ID":"ed370895-f652-499d-8382-69fb05f78f03","Type":"ContainerStarted","Data":"0c23f54bbe4b821652dd4ac5b83934b0905a9bfb198769aaee47c935549afdec"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.356309 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" event={"ID":"ed370895-f652-499d-8382-69fb05f78f03","Type":"ContainerStarted","Data":"d40270137f671c1e1373b1d229ef98158e72d966c6f1e742599bab3976c39304"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.420293 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" event={"ID":"ec0bf78c-4e35-4129-8976-c8d805991843","Type":"ContainerStarted","Data":"ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.420432 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" podStartSLOduration=133.42041074 podStartE2EDuration="2m13.42041074s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:24.233379734 +0000 UTC m=+152.298795342" watchObservedRunningTime="2025-12-11 08:39:24.42041074 +0000 UTC m=+152.485826348" Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.421994 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.427920 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:24 crc kubenswrapper[4629]: E1211 08:39:24.430071 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:24.930038667 +0000 UTC m=+152.995454445 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.455725 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.502123 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" podStartSLOduration=132.50209872 podStartE2EDuration="2m12.50209872s" podCreationTimestamp="2025-12-11 08:37:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:24.499229529 +0000 UTC m=+152.564645137" watchObservedRunningTime="2025-12-11 08:39:24.50209872 +0000 UTC m=+152.567514328" Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.514404 4629 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-4jtwf container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.16:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.514499 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" podUID="449eb3bb-7054-452d-85d6-558aa5efa682" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.16:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.516445 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" event={"ID":"cca13586-fe9e-43e9-acb8-c3ff573ff1db","Type":"ContainerStarted","Data":"9dcc93a343909c10725c3c5745be96257d61d5c8698dcee9f3e7d59ec9eb6bb7"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.527266 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rzxln" event={"ID":"92906dea-785c-45c4-95fc-9e6ed15fc609","Type":"ContainerStarted","Data":"ef5c91ec13842710c6ef416f5b6871692b7d29c2ab3a22ef60fd703f164eee7f"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.534126 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mbkcz" event={"ID":"bfbd1947-e471-47dd-8c36-916c28cfaf79","Type":"ContainerStarted","Data":"04db042bd94727ff2c3bce5b6dd7eb779e4d00488cfe731b85ccae0a692eafdf"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.536427 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:24 crc kubenswrapper[4629]: E1211 08:39:24.537129 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:25.037113969 +0000 UTC m=+153.102529577 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.541861 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" event={"ID":"e84237c3-8fcc-4a25-9d54-2422dcc628c5","Type":"ContainerStarted","Data":"35a9408e93ad64356c58570fa9de8669469e300064772f2b2e68e6c983e705db"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.550752 4629 generic.go:334] "Generic (PLEG): container finished" podID="0c0faec1-d9b4-49c4-8595-381e5d0c4a56" containerID="779f6eec169a4630b722eebc030de8bcae3a0ef9ecbb17daa6138388a8685a09" exitCode=0 Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.550867 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" event={"ID":"0c0faec1-d9b4-49c4-8595-381e5d0c4a56","Type":"ContainerDied","Data":"779f6eec169a4630b722eebc030de8bcae3a0ef9ecbb17daa6138388a8685a09"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.555261 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5ccrx" event={"ID":"9311bda6-d765-4970-9749-3fc2b826b0a3","Type":"ContainerStarted","Data":"75f6e59803803d3694c90e37f00ac135c5e166688280ac5f32a83c86e14f4e76"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.556478 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" event={"ID":"a33a0903-3d69-4017-b274-4f188c6b89c8","Type":"ContainerStarted","Data":"e2be5da444e5b9584427182850237653b2e5a68ffc276ff1cbc2152f7bdba1ea"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.558141 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv" event={"ID":"28a9a7b5-a9ca-4ac4-bef4-da3715a27b67","Type":"ContainerStarted","Data":"766c53cf6c71f5f09d8c9407d46f1f8965eead1103a3c23b906edce19f3db60e"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.558167 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv" event={"ID":"28a9a7b5-a9ca-4ac4-bef4-da3715a27b67","Type":"ContainerStarted","Data":"d9b921a79d812b374a3dcfa1347f2681780299c07f0ad8e97b301109b56f73c0"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.559933 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" event={"ID":"8bf66625-0e08-4c99-9633-85558c1ce12f","Type":"ContainerStarted","Data":"d34a1dcb28e6a6712d5c7ed4025f38d9356bafeb3d0669c9915f887c0ccedead"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.560806 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-wbrg7" event={"ID":"83f7a7a2-12d8-43b0-9a3a-3926d3f0092e","Type":"ContainerStarted","Data":"67f8c3b94c2a5090802aedc686c62ac595efe4e254a0d223c57f5db40befd83b"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.560824 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-wbrg7" event={"ID":"83f7a7a2-12d8-43b0-9a3a-3926d3f0092e","Type":"ContainerStarted","Data":"0f8653db279ca4a2030810214b80d37d17dc95fb3c9de34d32f0c7df6490197a"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.561493 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-wbrg7" Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.562741 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" event={"ID":"d2d4978b-296a-4177-955e-d367860c19ab","Type":"ContainerStarted","Data":"70f6cda1c0ef0ca58190f786f043f189915a25a539df3ba820bc61b6c53d4c04"} Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.569954 4629 patch_prober.go:28] interesting pod/downloads-7954f5f757-wbrg7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.570002 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wbrg7" podUID="83f7a7a2-12d8-43b0-9a3a-3926d3f0092e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.571198 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-npk4p" podStartSLOduration=133.571188538 podStartE2EDuration="2m13.571188538s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:24.564481323 +0000 UTC m=+152.629896931" watchObservedRunningTime="2025-12-11 08:39:24.571188538 +0000 UTC m=+152.636604146" Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.651241 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:24 crc kubenswrapper[4629]: E1211 08:39:24.658370 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:25.158349413 +0000 UTC m=+153.223765021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.751328 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5bwv" podStartSLOduration=133.751303493 podStartE2EDuration="2m13.751303493s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:24.748916747 +0000 UTC m=+152.814332365" watchObservedRunningTime="2025-12-11 08:39:24.751303493 +0000 UTC m=+152.816719101" Dec 11 08:39:24 crc kubenswrapper[4629]: I1211 08:39:24.755038 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:24 crc kubenswrapper[4629]: E1211 08:39:24.755571 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:25.255552109 +0000 UTC m=+153.320967717 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:24.954501 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:25 crc kubenswrapper[4629]: E1211 08:39:24.954831 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:25.454814456 +0000 UTC m=+153.520230064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:24.961301 4629 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7gkf5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:24.961406 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" podUID="ab34fa3c-8110-4868-95cf-2ec4ec5c3ceb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.31:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.083157 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:25 crc kubenswrapper[4629]: E1211 08:39:25.083450 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:25.583438276 +0000 UTC m=+153.648853884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.201589 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:25 crc kubenswrapper[4629]: E1211 08:39:25.202298 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:25.702271374 +0000 UTC m=+153.767686982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.307824 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rrmqz" podStartSLOduration=134.307807346 podStartE2EDuration="2m14.307807346s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:25.306436171 +0000 UTC m=+153.371851779" watchObservedRunningTime="2025-12-11 08:39:25.307807346 +0000 UTC m=+153.373222954" Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.309669 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:25 crc kubenswrapper[4629]: E1211 08:39:25.310015 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:25.809997775 +0000 UTC m=+153.875413393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.339885 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-wbrg7" podStartSLOduration=134.339839739 podStartE2EDuration="2m14.339839739s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:25.339017293 +0000 UTC m=+153.404432901" watchObservedRunningTime="2025-12-11 08:39:25.339839739 +0000 UTC m=+153.405255347" Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.360406 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:25 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:25 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:25 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.360502 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.411325 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:25 crc kubenswrapper[4629]: E1211 08:39:25.412096 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:25.912051697 +0000 UTC m=+153.977467305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.516227 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:25 crc kubenswrapper[4629]: E1211 08:39:25.516760 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:26.016741662 +0000 UTC m=+154.082157270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.535562 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-8q9kf" Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.618245 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:25 crc kubenswrapper[4629]: E1211 08:39:25.618684 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:26.118655798 +0000 UTC m=+154.184071406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.661028 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" event={"ID":"99df09c3-eae2-4db8-a82e-a7e1d1c462e5","Type":"ContainerStarted","Data":"0b6c51b698e52977d4507e1d28e74bed00792b9aef1f5146d036958ad18dc369"} Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.685586 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" event={"ID":"d44b0b79-360a-4d50-abb1-ef1ded04c565","Type":"ContainerStarted","Data":"f2e628e91d208796fc4f082b95d1297bba9f3d0f4360ead82c365c6b81870014"} Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.720606 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:25 crc kubenswrapper[4629]: E1211 08:39:25.721074 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:26.22106179 +0000 UTC m=+154.286477388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.746409 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" event={"ID":"0463664b-27fa-42ed-8369-442d93fbdcb5","Type":"ContainerStarted","Data":"6ba4396c076efe9795e8f3529764c9d3ac1a87b1d16a01bcfc2fb1f069f01037"} Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.785860 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" event={"ID":"beac9023-3506-4a56-9ac8-667172e2760d","Type":"ContainerStarted","Data":"4dbec18c213d56446990763f8400e354512764299ca5fea9809dbcf1a7ecda6f"} Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.786349 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" podStartSLOduration=134.786331566 podStartE2EDuration="2m14.786331566s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:25.784295401 +0000 UTC m=+153.849711009" watchObservedRunningTime="2025-12-11 08:39:25.786331566 +0000 UTC m=+153.851747174" Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.787318 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.796417 4629 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-v94fb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.796490 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" podUID="beac9023-3506-4a56-9ac8-667172e2760d" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.827569 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mbkcz" event={"ID":"bfbd1947-e471-47dd-8c36-916c28cfaf79","Type":"ContainerStarted","Data":"ee8e53c1fc84591431ecd1c01f197e6c53d2a01b9c7b41fdcb8d90064e8e692f"} Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.833548 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:25 crc kubenswrapper[4629]: E1211 08:39:25.834686 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:26.334664081 +0000 UTC m=+154.400079759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.865959 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fmmgm" event={"ID":"3aa8cf22-2a61-42f4-ac16-6d47ecc1eaa8","Type":"ContainerStarted","Data":"483a87e869185d5369ef154a7ffc721ba31fd9d363a685f8cc54efe3d6047ee8"} Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.878239 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222" event={"ID":"aedfd32e-f597-4753-8fbc-824b562b9c77","Type":"ContainerStarted","Data":"6210337d95adfb9cef4cb8914fc4f47ab06889d1890d46df74adde3d3b192877"} Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.898632 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" podStartSLOduration=134.898615014 podStartE2EDuration="2m14.898615014s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:25.896803137 +0000 UTC m=+153.962218745" watchObservedRunningTime="2025-12-11 08:39:25.898615014 +0000 UTC m=+153.964030622" Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.907452 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-n5p54" event={"ID":"6dcbb152-ea3e-4db5-aa09-d60e3768a99c","Type":"ContainerStarted","Data":"ee342ea27dc3959f461c23b4d17e8824b14b210248cc6c64a40fbd0787c60431"} Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.926066 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" event={"ID":"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc","Type":"ContainerStarted","Data":"26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed"} Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.926725 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.931035 4629 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rzqjs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.931108 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" podUID="128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.934602 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:25 crc kubenswrapper[4629]: E1211 08:39:25.937581 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:26.437566339 +0000 UTC m=+154.502981947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.960058 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6qgqv" event={"ID":"07febd9a-7813-4722-af4d-9b165be26662","Type":"ContainerStarted","Data":"f26976450a2da95e91f79fcba34ef50f88acdeec0bc6d45fa3b194333c8c0252"} Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.973438 4629 generic.go:334] "Generic (PLEG): container finished" podID="c34dd798-968c-4df8-bd03-8308d64fcd8c" containerID="a12e2a6ce9fe4354219af746136f01e6a8d1296c15ce6f79446c1af1bcc297ac" exitCode=0 Dec 11 08:39:25 crc kubenswrapper[4629]: I1211 08:39:25.973575 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" event={"ID":"c34dd798-968c-4df8-bd03-8308d64fcd8c","Type":"ContainerDied","Data":"a12e2a6ce9fe4354219af746136f01e6a8d1296c15ce6f79446c1af1bcc297ac"} Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.012892 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" event={"ID":"98e995fc-e7f7-48ef-bf7a-5e42bd9f8a55","Type":"ContainerStarted","Data":"04eba6a189f961b8b147b3613229d19c6506eb859fc1843579e11f4a476e9b4e"} Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.019046 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" event={"ID":"3efc8f47-f897-4072-b1ea-da866707146c","Type":"ContainerStarted","Data":"7c64a86b7514f4ef22f2e88aeeb96d6427e4ee60a4cfa4f4393b4cd4ac864d48"} Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.020036 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.027997 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" event={"ID":"8bf66625-0e08-4c99-9633-85558c1ce12f","Type":"ContainerStarted","Data":"c0fcb3ec1705c9f3ed483085dbcd57afb386da6dafed6a2ead8a92375982c747"} Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.038984 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:26 crc kubenswrapper[4629]: E1211 08:39:26.039438 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:26.539416774 +0000 UTC m=+154.604832382 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.068415 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" event={"ID":"9e92bcf9-0313-43ca-b2ae-fbdd9586aea2","Type":"ContainerStarted","Data":"85053b1b573373ffe909a6af6efa244f873fd89cb2ca29ec81eec47038a7dd4a"} Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.074126 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.110704 4629 patch_prober.go:28] interesting pod/downloads-7954f5f757-wbrg7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.110760 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wbrg7" podUID="83f7a7a2-12d8-43b0-9a3a-3926d3f0092e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.127697 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.140736 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:26 crc kubenswrapper[4629]: E1211 08:39:26.143983 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:26.643963294 +0000 UTC m=+154.709378962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.158358 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-fmmgm" podStartSLOduration=135.158339133 podStartE2EDuration="2m15.158339133s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:26.024018061 +0000 UTC m=+154.089433669" watchObservedRunningTime="2025-12-11 08:39:26.158339133 +0000 UTC m=+154.223754731" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.241800 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:26 crc kubenswrapper[4629]: E1211 08:39:26.243486 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:26.743467464 +0000 UTC m=+154.808883082 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.311349 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m9222" podStartSLOduration=135.311330192 podStartE2EDuration="2m15.311330192s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:26.282418098 +0000 UTC m=+154.347833706" watchObservedRunningTime="2025-12-11 08:39:26.311330192 +0000 UTC m=+154.376745800" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.312265 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mbkcz" podStartSLOduration=135.312259032 podStartE2EDuration="2m15.312259032s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:26.158948703 +0000 UTC m=+154.224364311" watchObservedRunningTime="2025-12-11 08:39:26.312259032 +0000 UTC m=+154.377674660" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.349645 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:26 crc kubenswrapper[4629]: E1211 08:39:26.350383 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:26.85037115 +0000 UTC m=+154.915786758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.356104 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:26 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:26 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:26 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.356151 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.450906 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:26 crc kubenswrapper[4629]: E1211 08:39:26.451194 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:26.951179441 +0000 UTC m=+155.016595039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.454416 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" podStartSLOduration=135.454394553 podStartE2EDuration="2m15.454394553s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:26.406116091 +0000 UTC m=+154.471531699" watchObservedRunningTime="2025-12-11 08:39:26.454394553 +0000 UTC m=+154.519810161" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.560206 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:26 crc kubenswrapper[4629]: E1211 08:39:26.560541 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:27.060529474 +0000 UTC m=+155.125945082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.580237 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-5ccrx" podStartSLOduration=135.580218444 podStartE2EDuration="2m15.580218444s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:26.459793816 +0000 UTC m=+154.525209424" watchObservedRunningTime="2025-12-11 08:39:26.580218444 +0000 UTC m=+154.645634052" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.660892 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:26 crc kubenswrapper[4629]: E1211 08:39:26.661245 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:27.161230822 +0000 UTC m=+155.226646430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.673527 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rw8lc" podStartSLOduration=135.673508525 podStartE2EDuration="2m15.673508525s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:26.672381949 +0000 UTC m=+154.737797547" watchObservedRunningTime="2025-12-11 08:39:26.673508525 +0000 UTC m=+154.738924143" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.675344 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lr8q8" podStartSLOduration=135.675331103 podStartE2EDuration="2m15.675331103s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:26.595323077 +0000 UTC m=+154.660738695" watchObservedRunningTime="2025-12-11 08:39:26.675331103 +0000 UTC m=+154.740746711" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.762267 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:26 crc kubenswrapper[4629]: E1211 08:39:26.762655 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:27.262637943 +0000 UTC m=+155.328053551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.833027 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" podStartSLOduration=135.833012672 podStartE2EDuration="2m15.833012672s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:26.828645033 +0000 UTC m=+154.894060641" watchObservedRunningTime="2025-12-11 08:39:26.833012672 +0000 UTC m=+154.898428280" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.863660 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:26 crc kubenswrapper[4629]: E1211 08:39:26.864086 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:27.364071124 +0000 UTC m=+155.429486732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.895285 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7gkf5" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.903254 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-x9fpk" podStartSLOduration=135.903237046 podStartE2EDuration="2m15.903237046s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:26.903127263 +0000 UTC m=+154.968542881" watchObservedRunningTime="2025-12-11 08:39:26.903237046 +0000 UTC m=+154.968652654" Dec 11 08:39:26 crc kubenswrapper[4629]: I1211 08:39:26.967293 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:26 crc kubenswrapper[4629]: E1211 08:39:26.967723 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:27.467707036 +0000 UTC m=+155.533122644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.068317 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.068672 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:27.568655552 +0000 UTC m=+155.634071160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.131448 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" event={"ID":"d957c3d0-bff4-4c92-82ea-3900ea7efa06","Type":"ContainerStarted","Data":"c8d58bb7dbb03331684de56cd12a5d9a12ae7a89bfcd1f5648602de74c1111ac"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.132330 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.137353 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rzxln" event={"ID":"92906dea-785c-45c4-95fc-9e6ed15fc609","Type":"ContainerStarted","Data":"22ebe50f235fdfb1410166d68291c8c9f5f8d66315034ebfd704482f353de070"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.138894 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5p2r" event={"ID":"8bf66625-0e08-4c99-9633-85558c1ce12f","Type":"ContainerStarted","Data":"d1a59e4e8b8d5a4c414d37ce4d3cc4f627daab3d9f7ff611b36f976500d75027"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.142508 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" event={"ID":"d2d4978b-296a-4177-955e-d367860c19ab","Type":"ContainerStarted","Data":"4d200f1bb5e1e503bd40991a1b7ae07da27b15467376506e38e25379d3be6303"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.148810 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mbkcz" event={"ID":"bfbd1947-e471-47dd-8c36-916c28cfaf79","Type":"ContainerStarted","Data":"a43d81172d4be6e089fa6b5812df78ca7937335078d45527650b9a51c4b12ff0"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.150995 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5ccrx" event={"ID":"9311bda6-d765-4970-9749-3fc2b826b0a3","Type":"ContainerStarted","Data":"8622c59bd210e2ea77f57986eea4cd33a1d5ea89d1848d6570c850a14f28107a"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.153080 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" event={"ID":"c34dd798-968c-4df8-bd03-8308d64fcd8c","Type":"ContainerStarted","Data":"966319b048584c6a19c2690e45a1050c117c4bff1032effd70f980d866d4e191"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.170347 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-n5p54" event={"ID":"6dcbb152-ea3e-4db5-aa09-d60e3768a99c","Type":"ContainerStarted","Data":"530df86d358bb1da59b1eb7def011afb814f1fb0a3e72fc49eced0adb9c33d7b"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.170973 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.172531 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.172857 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:27.672832801 +0000 UTC m=+155.738248409 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.178440 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" event={"ID":"e84237c3-8fcc-4a25-9d54-2422dcc628c5","Type":"ContainerStarted","Data":"cdc83d82e754ac43e29b3de33ced907dc50d700f295e5ca9f3c968d53561baa3"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.188183 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" event={"ID":"0c0faec1-d9b4-49c4-8595-381e5d0c4a56","Type":"ContainerStarted","Data":"40661d22b8989b0694def0872d9f2ca6dc0a01d26e726b1d1daeed889023e784"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.196297 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" event={"ID":"a33a0903-3d69-4017-b274-4f188c6b89c8","Type":"ContainerStarted","Data":"12de80fedc90741911c39cd2426029d4e50611c7d66476458c1316d4e5f460aa"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.196338 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" event={"ID":"a33a0903-3d69-4017-b274-4f188c6b89c8","Type":"ContainerStarted","Data":"f135af64ef29a8948ce13b89b6de41c0951017f6821d4acfdf0d5370d6d4f205"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.234559 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" event={"ID":"0463664b-27fa-42ed-8369-442d93fbdcb5","Type":"ContainerStarted","Data":"08222b7144e85df9fa4261764662c1b67c949df339b9f1e7c6d1026465879346"} Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.237239 4629 patch_prober.go:28] interesting pod/downloads-7954f5f757-wbrg7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.237279 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wbrg7" podUID="83f7a7a2-12d8-43b0-9a3a-3926d3f0092e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.237560 4629 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rzqjs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.237586 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" podUID="128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.252179 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v94fb" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.273509 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.273663 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:27.773640822 +0000 UTC m=+155.839056430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.273804 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.275172 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:27.77516461 +0000 UTC m=+155.840580218 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.359159 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:27 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:27 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:27 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.359223 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.375013 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.375699 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:27.875681242 +0000 UTC m=+155.941096850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.414895 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-n5p54" podStartSLOduration=11.414825583 podStartE2EDuration="11.414825583s" podCreationTimestamp="2025-12-11 08:39:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:27.40440636 +0000 UTC m=+155.469821968" watchObservedRunningTime="2025-12-11 08:39:27.414825583 +0000 UTC m=+155.480241191" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.415146 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" podStartSLOduration=136.415141713 podStartE2EDuration="2m16.415141713s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:27.295986386 +0000 UTC m=+155.361401994" watchObservedRunningTime="2025-12-11 08:39:27.415141713 +0000 UTC m=+155.480557311" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.477950 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.478259 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:27.97824768 +0000 UTC m=+156.043663288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.508295 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" podStartSLOduration=136.508275149 podStartE2EDuration="2m16.508275149s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:27.502305478 +0000 UTC m=+155.567721086" watchObservedRunningTime="2025-12-11 08:39:27.508275149 +0000 UTC m=+155.573690757" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.579260 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.579537 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.079523566 +0000 UTC m=+156.144939174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.618112 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" podStartSLOduration=135.618098569 podStartE2EDuration="2m15.618098569s" podCreationTimestamp="2025-12-11 08:37:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:27.577446289 +0000 UTC m=+155.642861897" watchObservedRunningTime="2025-12-11 08:39:27.618098569 +0000 UTC m=+155.683514177" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.681288 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.681589 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.181578927 +0000 UTC m=+156.246994535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.688666 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-rzxln" podStartSLOduration=11.688646792 podStartE2EDuration="11.688646792s" podCreationTimestamp="2025-12-11 08:39:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:27.620589778 +0000 UTC m=+155.686005386" watchObservedRunningTime="2025-12-11 08:39:27.688646792 +0000 UTC m=+155.754062400" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.781799 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.781974 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.281948624 +0000 UTC m=+156.347364232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.782060 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.782345 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.282335027 +0000 UTC m=+156.347750635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.882826 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.883009 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.382977102 +0000 UTC m=+156.448392720 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.883120 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.883441 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.383431327 +0000 UTC m=+156.448846935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.910066 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8v64" podStartSLOduration=136.910043117 podStartE2EDuration="2m16.910043117s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:27.771225712 +0000 UTC m=+155.836641320" watchObservedRunningTime="2025-12-11 08:39:27.910043117 +0000 UTC m=+155.975458725" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.984070 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:27 crc kubenswrapper[4629]: E1211 08:39:27.984461 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.484435135 +0000 UTC m=+156.549850743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.993021 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-947kd" podStartSLOduration=136.993005399 podStartE2EDuration="2m16.993005399s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:27.910538893 +0000 UTC m=+155.975954501" watchObservedRunningTime="2025-12-11 08:39:27.993005399 +0000 UTC m=+156.058421007" Dec 11 08:39:27 crc kubenswrapper[4629]: I1211 08:39:27.993246 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pvsn8" podStartSLOduration=136.993241836 podStartE2EDuration="2m16.993241836s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:27.991870482 +0000 UTC m=+156.057286090" watchObservedRunningTime="2025-12-11 08:39:27.993241836 +0000 UTC m=+156.058657444" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.085398 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:28 crc kubenswrapper[4629]: E1211 08:39:28.085799 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.585786083 +0000 UTC m=+156.651201691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.186894 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:28 crc kubenswrapper[4629]: E1211 08:39:28.187078 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.687049588 +0000 UTC m=+156.752465196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.187407 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:28 crc kubenswrapper[4629]: E1211 08:39:28.187676 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.687664618 +0000 UTC m=+156.753080226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.241222 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" event={"ID":"0c0faec1-d9b4-49c4-8595-381e5d0c4a56","Type":"ContainerStarted","Data":"0edcb5efdae35b8cc78c9922d0d76bae46c6c0632f4663010b95ece1c6945c0f"} Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.243767 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" event={"ID":"d44b0b79-360a-4d50-abb1-ef1ded04c565","Type":"ContainerStarted","Data":"43b336fae5243b5c517b25bcd76623d102efe7e5232778cb9ef962a01d79728b"} Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.244124 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.288544 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:28 crc kubenswrapper[4629]: E1211 08:39:28.288922 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.788863542 +0000 UTC m=+156.854279140 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.357366 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:28 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:28 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:28 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.357432 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.364898 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" podStartSLOduration=137.364859931 podStartE2EDuration="2m17.364859931s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:28.293453768 +0000 UTC m=+156.358869376" watchObservedRunningTime="2025-12-11 08:39:28.364859931 +0000 UTC m=+156.430275539" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.366556 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n2r4l"] Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.367889 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.372244 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.390744 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:28 crc kubenswrapper[4629]: E1211 08:39:28.395499 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.895483629 +0000 UTC m=+156.960899237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.426370 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n2r4l"] Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.492453 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:28 crc kubenswrapper[4629]: E1211 08:39:28.492658 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.992629373 +0000 UTC m=+157.058044981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.492749 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfb1548-f379-4fb3-ba39-b138585900b4-utilities\") pod \"certified-operators-n2r4l\" (UID: \"ccfb1548-f379-4fb3-ba39-b138585900b4\") " pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.492777 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfb1548-f379-4fb3-ba39-b138585900b4-catalog-content\") pod \"certified-operators-n2r4l\" (UID: \"ccfb1548-f379-4fb3-ba39-b138585900b4\") " pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.492814 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.492831 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q47jn\" (UniqueName: \"kubernetes.io/projected/ccfb1548-f379-4fb3-ba39-b138585900b4-kube-api-access-q47jn\") pod \"certified-operators-n2r4l\" (UID: \"ccfb1548-f379-4fb3-ba39-b138585900b4\") " pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:39:28 crc kubenswrapper[4629]: E1211 08:39:28.493154 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:28.99314478 +0000 UTC m=+157.058560388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.520486 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8f998"] Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.521405 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8f998" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.528682 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.593951 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.594126 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfb1548-f379-4fb3-ba39-b138585900b4-utilities\") pod \"certified-operators-n2r4l\" (UID: \"ccfb1548-f379-4fb3-ba39-b138585900b4\") " pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.594154 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfb1548-f379-4fb3-ba39-b138585900b4-catalog-content\") pod \"certified-operators-n2r4l\" (UID: \"ccfb1548-f379-4fb3-ba39-b138585900b4\") " pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.594191 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q47jn\" (UniqueName: \"kubernetes.io/projected/ccfb1548-f379-4fb3-ba39-b138585900b4-kube-api-access-q47jn\") pod \"certified-operators-n2r4l\" (UID: \"ccfb1548-f379-4fb3-ba39-b138585900b4\") " pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:39:28 crc kubenswrapper[4629]: E1211 08:39:28.594554 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:29.09453953 +0000 UTC m=+157.159955138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.594916 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfb1548-f379-4fb3-ba39-b138585900b4-utilities\") pod \"certified-operators-n2r4l\" (UID: \"ccfb1548-f379-4fb3-ba39-b138585900b4\") " pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.595121 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfb1548-f379-4fb3-ba39-b138585900b4-catalog-content\") pod \"certified-operators-n2r4l\" (UID: \"ccfb1548-f379-4fb3-ba39-b138585900b4\") " pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.616379 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q47jn\" (UniqueName: \"kubernetes.io/projected/ccfb1548-f379-4fb3-ba39-b138585900b4-kube-api-access-q47jn\") pod \"certified-operators-n2r4l\" (UID: \"ccfb1548-f379-4fb3-ba39-b138585900b4\") " pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.632550 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8f998"] Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.687328 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.702032 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c60b1455-35ba-47a0-a4d8-016c651e8c22-catalog-content\") pod \"community-operators-8f998\" (UID: \"c60b1455-35ba-47a0-a4d8-016c651e8c22\") " pod="openshift-marketplace/community-operators-8f998" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.702163 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c60b1455-35ba-47a0-a4d8-016c651e8c22-utilities\") pod \"community-operators-8f998\" (UID: \"c60b1455-35ba-47a0-a4d8-016c651e8c22\") " pod="openshift-marketplace/community-operators-8f998" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.702213 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm8t6\" (UniqueName: \"kubernetes.io/projected/c60b1455-35ba-47a0-a4d8-016c651e8c22-kube-api-access-tm8t6\") pod \"community-operators-8f998\" (UID: \"c60b1455-35ba-47a0-a4d8-016c651e8c22\") " pod="openshift-marketplace/community-operators-8f998" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.719963 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:28 crc kubenswrapper[4629]: E1211 08:39:28.720448 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:29.220432643 +0000 UTC m=+157.285848241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.719761 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kzp9g"] Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.725435 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.744250 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kzp9g"] Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.821326 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.821727 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c60b1455-35ba-47a0-a4d8-016c651e8c22-utilities\") pod \"community-operators-8f998\" (UID: \"c60b1455-35ba-47a0-a4d8-016c651e8c22\") " pod="openshift-marketplace/community-operators-8f998" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.821819 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm8t6\" (UniqueName: \"kubernetes.io/projected/c60b1455-35ba-47a0-a4d8-016c651e8c22-kube-api-access-tm8t6\") pod \"community-operators-8f998\" (UID: \"c60b1455-35ba-47a0-a4d8-016c651e8c22\") " pod="openshift-marketplace/community-operators-8f998" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.821959 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7adee9e0-edb2-4ea3-afa6-d5759440a79a-catalog-content\") pod \"certified-operators-kzp9g\" (UID: \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\") " pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.822043 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c60b1455-35ba-47a0-a4d8-016c651e8c22-catalog-content\") pod \"community-operators-8f998\" (UID: \"c60b1455-35ba-47a0-a4d8-016c651e8c22\") " pod="openshift-marketplace/community-operators-8f998" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.822073 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7adee9e0-edb2-4ea3-afa6-d5759440a79a-utilities\") pod \"certified-operators-kzp9g\" (UID: \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\") " pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.822110 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8q4g\" (UniqueName: \"kubernetes.io/projected/7adee9e0-edb2-4ea3-afa6-d5759440a79a-kube-api-access-c8q4g\") pod \"certified-operators-kzp9g\" (UID: \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\") " pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:39:28 crc kubenswrapper[4629]: E1211 08:39:28.822235 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:29.322214585 +0000 UTC m=+157.387630193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.823155 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c60b1455-35ba-47a0-a4d8-016c651e8c22-utilities\") pod \"community-operators-8f998\" (UID: \"c60b1455-35ba-47a0-a4d8-016c651e8c22\") " pod="openshift-marketplace/community-operators-8f998" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.827180 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c60b1455-35ba-47a0-a4d8-016c651e8c22-catalog-content\") pod \"community-operators-8f998\" (UID: \"c60b1455-35ba-47a0-a4d8-016c651e8c22\") " pod="openshift-marketplace/community-operators-8f998" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.844199 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.845310 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.846566 4629 patch_prober.go:28] interesting pod/console-f9d7485db-lhszp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.846609 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-lhszp" podUID="9d819e7a-6f7a-4f6e-bbba-4b334546f172" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.855791 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm8t6\" (UniqueName: \"kubernetes.io/projected/c60b1455-35ba-47a0-a4d8-016c651e8c22-kube-api-access-tm8t6\") pod \"community-operators-8f998\" (UID: \"c60b1455-35ba-47a0-a4d8-016c651e8c22\") " pod="openshift-marketplace/community-operators-8f998" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.905637 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-587qx"] Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.906971 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-587qx" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.924634 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7adee9e0-edb2-4ea3-afa6-d5759440a79a-utilities\") pod \"certified-operators-kzp9g\" (UID: \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\") " pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.924076 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7adee9e0-edb2-4ea3-afa6-d5759440a79a-utilities\") pod \"certified-operators-kzp9g\" (UID: \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\") " pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.924726 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-catalog-content\") pod \"community-operators-587qx\" (UID: \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\") " pod="openshift-marketplace/community-operators-587qx" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.924755 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwjpv\" (UniqueName: \"kubernetes.io/projected/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-kube-api-access-xwjpv\") pod \"community-operators-587qx\" (UID: \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\") " pod="openshift-marketplace/community-operators-587qx" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.924812 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8q4g\" (UniqueName: \"kubernetes.io/projected/7adee9e0-edb2-4ea3-afa6-d5759440a79a-kube-api-access-c8q4g\") pod \"certified-operators-kzp9g\" (UID: \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\") " pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.924965 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-utilities\") pod \"community-operators-587qx\" (UID: \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\") " pod="openshift-marketplace/community-operators-587qx" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.924993 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7adee9e0-edb2-4ea3-afa6-d5759440a79a-catalog-content\") pod \"certified-operators-kzp9g\" (UID: \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\") " pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.925067 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:28 crc kubenswrapper[4629]: E1211 08:39:28.925375 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:29.425360521 +0000 UTC m=+157.490776129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.926803 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7adee9e0-edb2-4ea3-afa6-d5759440a79a-catalog-content\") pod \"certified-operators-kzp9g\" (UID: \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\") " pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.941982 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-587qx"] Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.950779 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.950838 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:28 crc kubenswrapper[4629]: I1211 08:39:28.984374 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8q4g\" (UniqueName: \"kubernetes.io/projected/7adee9e0-edb2-4ea3-afa6-d5759440a79a-kube-api-access-c8q4g\") pod \"certified-operators-kzp9g\" (UID: \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\") " pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.007551 4629 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-xnsl5 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.14:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.007610 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" podUID="d957c3d0-bff4-4c92-82ea-3900ea7efa06" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.008969 4629 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-xnsl5 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.009017 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" podUID="d957c3d0-bff4-4c92-82ea-3900ea7efa06" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.026613 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.026859 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-catalog-content\") pod \"community-operators-587qx\" (UID: \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\") " pod="openshift-marketplace/community-operators-587qx" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.026911 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwjpv\" (UniqueName: \"kubernetes.io/projected/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-kube-api-access-xwjpv\") pod \"community-operators-587qx\" (UID: \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\") " pod="openshift-marketplace/community-operators-587qx" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.026984 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-utilities\") pod \"community-operators-587qx\" (UID: \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\") " pod="openshift-marketplace/community-operators-587qx" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.027495 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-utilities\") pod \"community-operators-587qx\" (UID: \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\") " pod="openshift-marketplace/community-operators-587qx" Dec 11 08:39:29 crc kubenswrapper[4629]: E1211 08:39:29.027589 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:29.527571527 +0000 UTC m=+157.592987125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.027901 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-catalog-content\") pod \"community-operators-587qx\" (UID: \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\") " pod="openshift-marketplace/community-operators-587qx" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.052701 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwjpv\" (UniqueName: \"kubernetes.io/projected/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-kube-api-access-xwjpv\") pod \"community-operators-587qx\" (UID: \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\") " pod="openshift-marketplace/community-operators-587qx" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.053103 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xnsl5" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.089170 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.130224 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:29 crc kubenswrapper[4629]: E1211 08:39:29.130502 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:29.630490235 +0000 UTC m=+157.695905843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.133948 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8f998" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.232475 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:29 crc kubenswrapper[4629]: E1211 08:39:29.232771 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:29.732756703 +0000 UTC m=+157.798172311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.248728 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-587qx" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.294594 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" event={"ID":"d44b0b79-360a-4d50-abb1-ef1ded04c565","Type":"ContainerStarted","Data":"8727b036e354214a962289654a02678b5b443236f2d8c9ba1760d05bfa0b2ff5"} Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.295942 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.296318 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.328565 4629 patch_prober.go:28] interesting pod/downloads-7954f5f757-wbrg7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.328943 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wbrg7" podUID="83f7a7a2-12d8-43b0-9a3a-3926d3f0092e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.329289 4629 patch_prober.go:28] interesting pod/downloads-7954f5f757-wbrg7 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.329316 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-wbrg7" podUID="83f7a7a2-12d8-43b0-9a3a-3926d3f0092e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.334685 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.346162 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:29 crc kubenswrapper[4629]: E1211 08:39:29.347143 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:29.847130198 +0000 UTC m=+157.912545806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.352437 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.370493 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:29 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:29 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:29 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.370545 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.441567 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n2r4l"] Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.442401 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:29 crc kubenswrapper[4629]: E1211 08:39:29.443697 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:29.943681554 +0000 UTC m=+158.009097162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.537695 4629 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rzqjs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.537747 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" podUID="128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.537797 4629 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rzqjs container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.537864 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" podUID="128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.547456 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:29 crc kubenswrapper[4629]: E1211 08:39:29.547948 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:30.047929374 +0000 UTC m=+158.113344982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.605617 4629 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.657629 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:29 crc kubenswrapper[4629]: E1211 08:39:29.658357 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:30.158343783 +0000 UTC m=+158.223759391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.759379 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:29 crc kubenswrapper[4629]: E1211 08:39:29.759668 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:30.25965639 +0000 UTC m=+158.325071998 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.861526 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:29 crc kubenswrapper[4629]: E1211 08:39:29.862088 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:30.362072412 +0000 UTC m=+158.427488020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.905908 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kzp9g"] Dec 11 08:39:29 crc kubenswrapper[4629]: I1211 08:39:29.965567 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:29 crc kubenswrapper[4629]: E1211 08:39:29.965855 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:30.465840338 +0000 UTC m=+158.531255946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.069564 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:30 crc kubenswrapper[4629]: E1211 08:39:30.069996 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:30.569974296 +0000 UTC m=+158.635389914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.159907 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-587qx"] Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.170973 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:30 crc kubenswrapper[4629]: E1211 08:39:30.171332 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:30.671320755 +0000 UTC m=+158.736736353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.273567 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:30 crc kubenswrapper[4629]: E1211 08:39:30.274096 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:39:30.774075578 +0000 UTC m=+158.839491186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.289372 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8f998"] Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.328621 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8f998" event={"ID":"c60b1455-35ba-47a0-a4d8-016c651e8c22","Type":"ContainerStarted","Data":"4db31d63398f2d468edb5ed52bab9df25d1ff358c21f88da55c3aa0cd6d360d8"} Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.352111 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" event={"ID":"d44b0b79-360a-4d50-abb1-ef1ded04c565","Type":"ContainerStarted","Data":"27aa8027fed87ae49d53a6f9240b089dc61c9ba8e7a081941a5e34fa7d9745ab"} Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.357689 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:30 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:30 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:30 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.357747 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.367535 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-587qx" event={"ID":"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366","Type":"ContainerStarted","Data":"7595dae8a7dd0b1dd85c49ec4a4aa379c86901871e38b9f5c62b568f369955e1"} Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.376044 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:30 crc kubenswrapper[4629]: E1211 08:39:30.376461 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:39:30.876446248 +0000 UTC m=+158.941861856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmklq" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.387976 4629 generic.go:334] "Generic (PLEG): container finished" podID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" containerID="c8393d9659441fe47b8496ca347c8a4618a1a881d2a3641de0be1a416f8943ff" exitCode=0 Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.388151 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzp9g" event={"ID":"7adee9e0-edb2-4ea3-afa6-d5759440a79a","Type":"ContainerDied","Data":"c8393d9659441fe47b8496ca347c8a4618a1a881d2a3641de0be1a416f8943ff"} Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.388207 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzp9g" event={"ID":"7adee9e0-edb2-4ea3-afa6-d5759440a79a","Type":"ContainerStarted","Data":"cc5fd0e5bb95ee30f7e59883c2192b5bf241ec02a35d75a0633097f59fbad50f"} Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.406645 4629 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-11T08:39:29.605646299Z","Handler":null,"Name":""} Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.414057 4629 generic.go:334] "Generic (PLEG): container finished" podID="ccfb1548-f379-4fb3-ba39-b138585900b4" containerID="4874bac62fb121e930fb70bf5fb575aff02b20d49f2dad3b77f63ce41896eef6" exitCode=0 Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.415260 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2r4l" event={"ID":"ccfb1548-f379-4fb3-ba39-b138585900b4","Type":"ContainerDied","Data":"4874bac62fb121e930fb70bf5fb575aff02b20d49f2dad3b77f63ce41896eef6"} Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.415320 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2r4l" event={"ID":"ccfb1548-f379-4fb3-ba39-b138585900b4","Type":"ContainerStarted","Data":"79acdfe15c099e03d4d026e4952cf15f081e5b8925575bff8652260494034bbf"} Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.424353 4629 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.437547 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5fhn" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.447757 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-wdcs7" podStartSLOduration=14.447719286 podStartE2EDuration="14.447719286s" podCreationTimestamp="2025-12-11 08:39:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:30.44280652 +0000 UTC m=+158.508222138" watchObservedRunningTime="2025-12-11 08:39:30.447719286 +0000 UTC m=+158.513134904" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.458357 4629 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.458406 4629 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.485456 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.656252 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.675097 4629 patch_prober.go:28] interesting pod/apiserver-76f77b778f-t7tg7 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 11 08:39:30 crc kubenswrapper[4629]: [+]log ok Dec 11 08:39:30 crc kubenswrapper[4629]: [+]etcd ok Dec 11 08:39:30 crc kubenswrapper[4629]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 11 08:39:30 crc kubenswrapper[4629]: [+]poststarthook/generic-apiserver-start-informers ok Dec 11 08:39:30 crc kubenswrapper[4629]: [+]poststarthook/max-in-flight-filter ok Dec 11 08:39:30 crc kubenswrapper[4629]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 11 08:39:30 crc kubenswrapper[4629]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 11 08:39:30 crc kubenswrapper[4629]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 11 08:39:30 crc kubenswrapper[4629]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 11 08:39:30 crc kubenswrapper[4629]: [+]poststarthook/project.openshift.io-projectcache ok Dec 11 08:39:30 crc kubenswrapper[4629]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 11 08:39:30 crc kubenswrapper[4629]: [+]poststarthook/openshift.io-startinformers ok Dec 11 08:39:30 crc kubenswrapper[4629]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 11 08:39:30 crc kubenswrapper[4629]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 11 08:39:30 crc kubenswrapper[4629]: livez check failed Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.675161 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" podUID="0c0faec1-d9b4-49c4-8595-381e5d0c4a56" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.691764 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.700564 4629 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.700625 4629 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.717923 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q8vkr"] Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.718821 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.729651 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.736127 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8vkr"] Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.792993 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28a0959a-1692-4c2e-83c3-54627cd1d53a-catalog-content\") pod \"redhat-marketplace-q8vkr\" (UID: \"28a0959a-1692-4c2e-83c3-54627cd1d53a\") " pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.793146 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kbnb\" (UniqueName: \"kubernetes.io/projected/28a0959a-1692-4c2e-83c3-54627cd1d53a-kube-api-access-4kbnb\") pod \"redhat-marketplace-q8vkr\" (UID: \"28a0959a-1692-4c2e-83c3-54627cd1d53a\") " pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.793193 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28a0959a-1692-4c2e-83c3-54627cd1d53a-utilities\") pod \"redhat-marketplace-q8vkr\" (UID: \"28a0959a-1692-4c2e-83c3-54627cd1d53a\") " pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.807110 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmklq\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.852424 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.894299 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kbnb\" (UniqueName: \"kubernetes.io/projected/28a0959a-1692-4c2e-83c3-54627cd1d53a-kube-api-access-4kbnb\") pod \"redhat-marketplace-q8vkr\" (UID: \"28a0959a-1692-4c2e-83c3-54627cd1d53a\") " pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.894361 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28a0959a-1692-4c2e-83c3-54627cd1d53a-utilities\") pod \"redhat-marketplace-q8vkr\" (UID: \"28a0959a-1692-4c2e-83c3-54627cd1d53a\") " pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.894483 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28a0959a-1692-4c2e-83c3-54627cd1d53a-catalog-content\") pod \"redhat-marketplace-q8vkr\" (UID: \"28a0959a-1692-4c2e-83c3-54627cd1d53a\") " pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.895194 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28a0959a-1692-4c2e-83c3-54627cd1d53a-catalog-content\") pod \"redhat-marketplace-q8vkr\" (UID: \"28a0959a-1692-4c2e-83c3-54627cd1d53a\") " pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.895453 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28a0959a-1692-4c2e-83c3-54627cd1d53a-utilities\") pod \"redhat-marketplace-q8vkr\" (UID: \"28a0959a-1692-4c2e-83c3-54627cd1d53a\") " pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:39:30 crc kubenswrapper[4629]: I1211 08:39:30.922812 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kbnb\" (UniqueName: \"kubernetes.io/projected/28a0959a-1692-4c2e-83c3-54627cd1d53a-kube-api-access-4kbnb\") pod \"redhat-marketplace-q8vkr\" (UID: \"28a0959a-1692-4c2e-83c3-54627cd1d53a\") " pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.042025 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.113507 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bwgwp"] Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.117531 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.155199 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwgwp"] Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.245582 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1451fb38-b06c-4b5c-bb43-8b571b8e3302-utilities\") pod \"redhat-marketplace-bwgwp\" (UID: \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\") " pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.245679 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zvjc\" (UniqueName: \"kubernetes.io/projected/1451fb38-b06c-4b5c-bb43-8b571b8e3302-kube-api-access-6zvjc\") pod \"redhat-marketplace-bwgwp\" (UID: \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\") " pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.245713 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1451fb38-b06c-4b5c-bb43-8b571b8e3302-catalog-content\") pod \"redhat-marketplace-bwgwp\" (UID: \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\") " pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.346413 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zvjc\" (UniqueName: \"kubernetes.io/projected/1451fb38-b06c-4b5c-bb43-8b571b8e3302-kube-api-access-6zvjc\") pod \"redhat-marketplace-bwgwp\" (UID: \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\") " pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.346452 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1451fb38-b06c-4b5c-bb43-8b571b8e3302-catalog-content\") pod \"redhat-marketplace-bwgwp\" (UID: \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\") " pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.346536 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1451fb38-b06c-4b5c-bb43-8b571b8e3302-utilities\") pod \"redhat-marketplace-bwgwp\" (UID: \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\") " pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.347044 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1451fb38-b06c-4b5c-bb43-8b571b8e3302-utilities\") pod \"redhat-marketplace-bwgwp\" (UID: \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\") " pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.347237 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1451fb38-b06c-4b5c-bb43-8b571b8e3302-catalog-content\") pod \"redhat-marketplace-bwgwp\" (UID: \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\") " pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.356424 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:31 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:31 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:31 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.356484 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.369975 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zvjc\" (UniqueName: \"kubernetes.io/projected/1451fb38-b06c-4b5c-bb43-8b571b8e3302-kube-api-access-6zvjc\") pod \"redhat-marketplace-bwgwp\" (UID: \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\") " pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.428368 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cmklq"] Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.429750 4629 generic.go:334] "Generic (PLEG): container finished" podID="c60b1455-35ba-47a0-a4d8-016c651e8c22" containerID="7cf74e82ff2a57d0552522b860fbbe85844e94d007f6da653dec65d6187369e1" exitCode=0 Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.429816 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8f998" event={"ID":"c60b1455-35ba-47a0-a4d8-016c651e8c22","Type":"ContainerDied","Data":"7cf74e82ff2a57d0552522b860fbbe85844e94d007f6da653dec65d6187369e1"} Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.441824 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8vkr"] Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.443405 4629 generic.go:334] "Generic (PLEG): container finished" podID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" containerID="550b17b57296d5ebe8356dc21396dcd20fa5be9c4daff2e7600edb9886fe783f" exitCode=0 Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.443525 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-587qx" event={"ID":"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366","Type":"ContainerDied","Data":"550b17b57296d5ebe8356dc21396dcd20fa5be9c4daff2e7600edb9886fe783f"} Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.458433 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.514037 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-69vgp"] Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.515184 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.520276 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.523149 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-69vgp"] Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.659428 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwdfq\" (UniqueName: \"kubernetes.io/projected/ecaf1148-6695-4cdf-9ffc-06cab2716241-kube-api-access-wwdfq\") pod \"redhat-operators-69vgp\" (UID: \"ecaf1148-6695-4cdf-9ffc-06cab2716241\") " pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.659519 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecaf1148-6695-4cdf-9ffc-06cab2716241-utilities\") pod \"redhat-operators-69vgp\" (UID: \"ecaf1148-6695-4cdf-9ffc-06cab2716241\") " pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.659785 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecaf1148-6695-4cdf-9ffc-06cab2716241-catalog-content\") pod \"redhat-operators-69vgp\" (UID: \"ecaf1148-6695-4cdf-9ffc-06cab2716241\") " pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.717266 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c6kzc"] Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.719744 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.733331 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6kzc"] Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.761659 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwdfq\" (UniqueName: \"kubernetes.io/projected/ecaf1148-6695-4cdf-9ffc-06cab2716241-kube-api-access-wwdfq\") pod \"redhat-operators-69vgp\" (UID: \"ecaf1148-6695-4cdf-9ffc-06cab2716241\") " pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.761774 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecaf1148-6695-4cdf-9ffc-06cab2716241-utilities\") pod \"redhat-operators-69vgp\" (UID: \"ecaf1148-6695-4cdf-9ffc-06cab2716241\") " pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.761824 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecaf1148-6695-4cdf-9ffc-06cab2716241-catalog-content\") pod \"redhat-operators-69vgp\" (UID: \"ecaf1148-6695-4cdf-9ffc-06cab2716241\") " pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.762601 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecaf1148-6695-4cdf-9ffc-06cab2716241-utilities\") pod \"redhat-operators-69vgp\" (UID: \"ecaf1148-6695-4cdf-9ffc-06cab2716241\") " pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.762800 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecaf1148-6695-4cdf-9ffc-06cab2716241-catalog-content\") pod \"redhat-operators-69vgp\" (UID: \"ecaf1148-6695-4cdf-9ffc-06cab2716241\") " pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.826990 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwdfq\" (UniqueName: \"kubernetes.io/projected/ecaf1148-6695-4cdf-9ffc-06cab2716241-kube-api-access-wwdfq\") pod \"redhat-operators-69vgp\" (UID: \"ecaf1148-6695-4cdf-9ffc-06cab2716241\") " pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.835532 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.865996 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-catalog-content\") pod \"redhat-operators-c6kzc\" (UID: \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\") " pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.866110 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-utilities\") pod \"redhat-operators-c6kzc\" (UID: \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\") " pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.866135 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4qzq\" (UniqueName: \"kubernetes.io/projected/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-kube-api-access-v4qzq\") pod \"redhat-operators-c6kzc\" (UID: \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\") " pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.967823 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-catalog-content\") pod \"redhat-operators-c6kzc\" (UID: \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\") " pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.967923 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-utilities\") pod \"redhat-operators-c6kzc\" (UID: \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\") " pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.967974 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4qzq\" (UniqueName: \"kubernetes.io/projected/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-kube-api-access-v4qzq\") pod \"redhat-operators-c6kzc\" (UID: \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\") " pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.968439 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-catalog-content\") pod \"redhat-operators-c6kzc\" (UID: \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\") " pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:39:31 crc kubenswrapper[4629]: I1211 08:39:31.968570 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-utilities\") pod \"redhat-operators-c6kzc\" (UID: \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\") " pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.030433 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4qzq\" (UniqueName: \"kubernetes.io/projected/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-kube-api-access-v4qzq\") pod \"redhat-operators-c6kzc\" (UID: \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\") " pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.108513 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.178398 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwgwp"] Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.208898 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 11 08:39:32 crc kubenswrapper[4629]: W1211 08:39:32.216296 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1451fb38_b06c_4b5c_bb43_8b571b8e3302.slice/crio-806c7656870463a3776ec16ac0ec6b46c9e1d96a5293bfa9f314e634a8c968b7 WatchSource:0}: Error finding container 806c7656870463a3776ec16ac0ec6b46c9e1d96a5293bfa9f314e634a8c968b7: Status 404 returned error can't find the container with id 806c7656870463a3776ec16ac0ec6b46c9e1d96a5293bfa9f314e634a8c968b7 Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.367813 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:32 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:32 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:32 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.367894 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.502836 4629 generic.go:334] "Generic (PLEG): container finished" podID="28a0959a-1692-4c2e-83c3-54627cd1d53a" containerID="91023925122b1e130dd383ad68b646646d822b11bf66a2a5bf1c401966d0e34d" exitCode=0 Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.502942 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8vkr" event={"ID":"28a0959a-1692-4c2e-83c3-54627cd1d53a","Type":"ContainerDied","Data":"91023925122b1e130dd383ad68b646646d822b11bf66a2a5bf1c401966d0e34d"} Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.503002 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8vkr" event={"ID":"28a0959a-1692-4c2e-83c3-54627cd1d53a","Type":"ContainerStarted","Data":"83e827b82abeabe17d0310eb1d1e0fe1939bf0a065fb3dfd5b71b3d7b2949662"} Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.510572 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwgwp" event={"ID":"1451fb38-b06c-4b5c-bb43-8b571b8e3302","Type":"ContainerStarted","Data":"806c7656870463a3776ec16ac0ec6b46c9e1d96a5293bfa9f314e634a8c968b7"} Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.514018 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" event={"ID":"caa0ac44-0445-4d00-af37-e00e3c3d7d5d","Type":"ContainerStarted","Data":"9e68593d9990ab48b82fe2cb1e46fcd4f08708f413c80eef3d21cb3fe55e5342"} Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.514055 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" event={"ID":"caa0ac44-0445-4d00-af37-e00e3c3d7d5d","Type":"ContainerStarted","Data":"6c86c600db1b1ec60c55f57da586a335af3a81c694d79c75a575e5accc60ee30"} Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.514824 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.546022 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" podStartSLOduration=141.545991094 podStartE2EDuration="2m21.545991094s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:32.544643421 +0000 UTC m=+160.610059039" watchObservedRunningTime="2025-12-11 08:39:32.545991094 +0000 UTC m=+160.611406702" Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.616695 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-69vgp"] Dec 11 08:39:32 crc kubenswrapper[4629]: W1211 08:39:32.648776 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecaf1148_6695_4cdf_9ffc_06cab2716241.slice/crio-461194cbec6e758e224db7142633f02e346cefa646c63477ccbf6d0390d2b302 WatchSource:0}: Error finding container 461194cbec6e758e224db7142633f02e346cefa646c63477ccbf6d0390d2b302: Status 404 returned error can't find the container with id 461194cbec6e758e224db7142633f02e346cefa646c63477ccbf6d0390d2b302 Dec 11 08:39:32 crc kubenswrapper[4629]: I1211 08:39:32.869642 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6kzc"] Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.158900 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.160008 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.163691 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.164660 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.165391 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.225295 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/44644f73-1584-404e-a605-01fb14c97921-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"44644f73-1584-404e-a605-01fb14c97921\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.225487 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44644f73-1584-404e-a605-01fb14c97921-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"44644f73-1584-404e-a605-01fb14c97921\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.328509 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/44644f73-1584-404e-a605-01fb14c97921-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"44644f73-1584-404e-a605-01fb14c97921\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.328623 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44644f73-1584-404e-a605-01fb14c97921-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"44644f73-1584-404e-a605-01fb14c97921\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.329084 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/44644f73-1584-404e-a605-01fb14c97921-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"44644f73-1584-404e-a605-01fb14c97921\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.347664 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44644f73-1584-404e-a605-01fb14c97921-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"44644f73-1584-404e-a605-01fb14c97921\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.357262 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:33 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:33 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:33 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.357317 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.487003 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.523831 4629 generic.go:334] "Generic (PLEG): container finished" podID="ecaf1148-6695-4cdf-9ffc-06cab2716241" containerID="d105d0b909583f816b453856eb5e3964ac57b8eaaad00fd43f86314c02c87e7e" exitCode=0 Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.523870 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69vgp" event={"ID":"ecaf1148-6695-4cdf-9ffc-06cab2716241","Type":"ContainerDied","Data":"d105d0b909583f816b453856eb5e3964ac57b8eaaad00fd43f86314c02c87e7e"} Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.523963 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69vgp" event={"ID":"ecaf1148-6695-4cdf-9ffc-06cab2716241","Type":"ContainerStarted","Data":"461194cbec6e758e224db7142633f02e346cefa646c63477ccbf6d0390d2b302"} Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.555030 4629 generic.go:334] "Generic (PLEG): container finished" podID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" containerID="d1958a466fe1da0a31395861397896ea391f7ad02b8de9a319a6c61a6d0f87f3" exitCode=0 Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.555121 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kzc" event={"ID":"69c02574-1a57-4d6c-aa9c-8a85c6acd80f","Type":"ContainerDied","Data":"d1958a466fe1da0a31395861397896ea391f7ad02b8de9a319a6c61a6d0f87f3"} Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.555147 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kzc" event={"ID":"69c02574-1a57-4d6c-aa9c-8a85c6acd80f","Type":"ContainerStarted","Data":"1f890ae78abd2f0d10dcc514e3acef8d79a872c2cc7b2fd00ae6ea021d0bf081"} Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.565039 4629 generic.go:334] "Generic (PLEG): container finished" podID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" containerID="98a218ce29b1195393ad792517bab9a04c942ef264cd61b8775bd22ea187978a" exitCode=0 Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.565731 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwgwp" event={"ID":"1451fb38-b06c-4b5c-bb43-8b571b8e3302","Type":"ContainerDied","Data":"98a218ce29b1195393ad792517bab9a04c942ef264cd61b8775bd22ea187978a"} Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.871294 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.954102 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:33 crc kubenswrapper[4629]: I1211 08:39:33.962187 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-t7tg7" Dec 11 08:39:34 crc kubenswrapper[4629]: I1211 08:39:34.145943 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:39:34 crc kubenswrapper[4629]: I1211 08:39:34.168819 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e996ab-2163-44ad-9ecc-052a44f0da4c-metrics-certs\") pod \"network-metrics-daemon-29fct\" (UID: \"13e996ab-2163-44ad-9ecc-052a44f0da4c\") " pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:39:34 crc kubenswrapper[4629]: I1211 08:39:34.357713 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:34 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:34 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:34 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:34 crc kubenswrapper[4629]: I1211 08:39:34.358110 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:34 crc kubenswrapper[4629]: I1211 08:39:34.411503 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-29fct" Dec 11 08:39:34 crc kubenswrapper[4629]: I1211 08:39:34.601304 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"44644f73-1584-404e-a605-01fb14c97921","Type":"ContainerStarted","Data":"d0456e9b9c85789c15baf510a752468830b342bc1bcb711eb32ab9c58af24f6e"} Dec 11 08:39:34 crc kubenswrapper[4629]: I1211 08:39:34.817365 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-29fct"] Dec 11 08:39:35 crc kubenswrapper[4629]: I1211 08:39:35.367090 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:35 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:35 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:35 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:35 crc kubenswrapper[4629]: I1211 08:39:35.367143 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:35 crc kubenswrapper[4629]: I1211 08:39:35.617956 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"44644f73-1584-404e-a605-01fb14c97921","Type":"ContainerStarted","Data":"b3c835e7e41bc96bf9b58b9821f094551700ef1f60616b90d13d7e9214cc449a"} Dec 11 08:39:35 crc kubenswrapper[4629]: I1211 08:39:35.624081 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-29fct" event={"ID":"13e996ab-2163-44ad-9ecc-052a44f0da4c","Type":"ContainerStarted","Data":"24119fecb1a3d0384dd2ee2548ad9cee0aa163dfad49f17afe82f80e0ad109a9"} Dec 11 08:39:35 crc kubenswrapper[4629]: I1211 08:39:35.624127 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-29fct" event={"ID":"13e996ab-2163-44ad-9ecc-052a44f0da4c","Type":"ContainerStarted","Data":"0ac4b7cbf2ed2631f4087029c2a49e8713295fa431b5cf8e960178d8bec3caa5"} Dec 11 08:39:35 crc kubenswrapper[4629]: I1211 08:39:35.652233 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.65220764 podStartE2EDuration="2.65220764s" podCreationTimestamp="2025-12-11 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:35.641955433 +0000 UTC m=+163.707371061" watchObservedRunningTime="2025-12-11 08:39:35.65220764 +0000 UTC m=+163.717623258" Dec 11 08:39:36 crc kubenswrapper[4629]: I1211 08:39:36.359811 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:36 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:36 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:36 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:36 crc kubenswrapper[4629]: I1211 08:39:36.360419 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:36 crc kubenswrapper[4629]: I1211 08:39:36.674986 4629 generic.go:334] "Generic (PLEG): container finished" podID="44644f73-1584-404e-a605-01fb14c97921" containerID="b3c835e7e41bc96bf9b58b9821f094551700ef1f60616b90d13d7e9214cc449a" exitCode=0 Dec 11 08:39:36 crc kubenswrapper[4629]: I1211 08:39:36.675073 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"44644f73-1584-404e-a605-01fb14c97921","Type":"ContainerDied","Data":"b3c835e7e41bc96bf9b58b9821f094551700ef1f60616b90d13d7e9214cc449a"} Dec 11 08:39:37 crc kubenswrapper[4629]: I1211 08:39:37.356682 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:37 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:37 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:37 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:37 crc kubenswrapper[4629]: I1211 08:39:37.356747 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:37 crc kubenswrapper[4629]: I1211 08:39:37.624924 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-n5p54" Dec 11 08:39:37 crc kubenswrapper[4629]: I1211 08:39:37.731531 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-29fct" event={"ID":"13e996ab-2163-44ad-9ecc-052a44f0da4c","Type":"ContainerStarted","Data":"3000a16ab0c36beb3bcebc97cd19e414197fd34b9a5ea40245147870b810eb7b"} Dec 11 08:39:37 crc kubenswrapper[4629]: I1211 08:39:37.748220 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-29fct" podStartSLOduration=146.748201366 podStartE2EDuration="2m26.748201366s" podCreationTimestamp="2025-12-11 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:37.747913796 +0000 UTC m=+165.813329414" watchObservedRunningTime="2025-12-11 08:39:37.748201366 +0000 UTC m=+165.813616974" Dec 11 08:39:37 crc kubenswrapper[4629]: I1211 08:39:37.854495 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 11 08:39:37 crc kubenswrapper[4629]: I1211 08:39:37.855434 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:39:37 crc kubenswrapper[4629]: I1211 08:39:37.860909 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 11 08:39:37 crc kubenswrapper[4629]: I1211 08:39:37.861282 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 11 08:39:37 crc kubenswrapper[4629]: I1211 08:39:37.874171 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 11 08:39:37 crc kubenswrapper[4629]: I1211 08:39:37.977780 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea2b9d9a-81ee-4120-be06-cc8a5684c765-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ea2b9d9a-81ee-4120-be06-cc8a5684c765\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:39:37 crc kubenswrapper[4629]: I1211 08:39:37.977837 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea2b9d9a-81ee-4120-be06-cc8a5684c765-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ea2b9d9a-81ee-4120-be06-cc8a5684c765\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.078985 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea2b9d9a-81ee-4120-be06-cc8a5684c765-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ea2b9d9a-81ee-4120-be06-cc8a5684c765\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.079058 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea2b9d9a-81ee-4120-be06-cc8a5684c765-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ea2b9d9a-81ee-4120-be06-cc8a5684c765\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.079154 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea2b9d9a-81ee-4120-be06-cc8a5684c765-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ea2b9d9a-81ee-4120-be06-cc8a5684c765\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.120812 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea2b9d9a-81ee-4120-be06-cc8a5684c765-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ea2b9d9a-81ee-4120-be06-cc8a5684c765\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.180225 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.195502 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.281444 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/44644f73-1584-404e-a605-01fb14c97921-kubelet-dir\") pod \"44644f73-1584-404e-a605-01fb14c97921\" (UID: \"44644f73-1584-404e-a605-01fb14c97921\") " Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.281916 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44644f73-1584-404e-a605-01fb14c97921-kube-api-access\") pod \"44644f73-1584-404e-a605-01fb14c97921\" (UID: \"44644f73-1584-404e-a605-01fb14c97921\") " Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.281594 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/44644f73-1584-404e-a605-01fb14c97921-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "44644f73-1584-404e-a605-01fb14c97921" (UID: "44644f73-1584-404e-a605-01fb14c97921"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.304637 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44644f73-1584-404e-a605-01fb14c97921-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "44644f73-1584-404e-a605-01fb14c97921" (UID: "44644f73-1584-404e-a605-01fb14c97921"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.353960 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:38 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:38 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:38 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.354028 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.383215 4629 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/44644f73-1584-404e-a605-01fb14c97921-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.383254 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44644f73-1584-404e-a605-01fb14c97921-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.762408 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.769538 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"44644f73-1584-404e-a605-01fb14c97921","Type":"ContainerDied","Data":"d0456e9b9c85789c15baf510a752468830b342bc1bcb711eb32ab9c58af24f6e"} Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.769593 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0456e9b9c85789c15baf510a752468830b342bc1bcb711eb32ab9c58af24f6e" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.769560 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.772152 4629 generic.go:334] "Generic (PLEG): container finished" podID="99df09c3-eae2-4db8-a82e-a7e1d1c462e5" containerID="0b6c51b698e52977d4507e1d28e74bed00792b9aef1f5146d036958ad18dc369" exitCode=0 Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.772206 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" event={"ID":"99df09c3-eae2-4db8-a82e-a7e1d1c462e5","Type":"ContainerDied","Data":"0b6c51b698e52977d4507e1d28e74bed00792b9aef1f5146d036958ad18dc369"} Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.844132 4629 patch_prober.go:28] interesting pod/console-f9d7485db-lhszp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 11 08:39:38 crc kubenswrapper[4629]: I1211 08:39:38.844201 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-lhszp" podUID="9d819e7a-6f7a-4f6e-bbba-4b334546f172" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 11 08:39:39 crc kubenswrapper[4629]: I1211 08:39:39.328573 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-wbrg7" Dec 11 08:39:39 crc kubenswrapper[4629]: I1211 08:39:39.362256 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:39 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:39 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:39 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:39 crc kubenswrapper[4629]: I1211 08:39:39.362346 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:39 crc kubenswrapper[4629]: I1211 08:39:39.539185 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:39:39 crc kubenswrapper[4629]: I1211 08:39:39.784007 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ea2b9d9a-81ee-4120-be06-cc8a5684c765","Type":"ContainerStarted","Data":"0c0040a3e942694d83f783465b1fe67bb874eaa1a62a04f0ef2d74835cbf5c80"} Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.306130 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.354022 4629 patch_prober.go:28] interesting pod/router-default-5444994796-r9x7m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:39:40 crc kubenswrapper[4629]: [-]has-synced failed: reason withheld Dec 11 08:39:40 crc kubenswrapper[4629]: [+]process-running ok Dec 11 08:39:40 crc kubenswrapper[4629]: healthz check failed Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.354121 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r9x7m" podUID="641fc441-d314-4b07-9fdf-98a81d21744f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.423121 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-secret-volume\") pod \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\" (UID: \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\") " Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.425682 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkqmk\" (UniqueName: \"kubernetes.io/projected/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-kube-api-access-tkqmk\") pod \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\" (UID: \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\") " Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.425864 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-config-volume\") pod \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\" (UID: \"99df09c3-eae2-4db8-a82e-a7e1d1c462e5\") " Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.426822 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-config-volume" (OuterVolumeSpecName: "config-volume") pod "99df09c3-eae2-4db8-a82e-a7e1d1c462e5" (UID: "99df09c3-eae2-4db8-a82e-a7e1d1c462e5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.430556 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "99df09c3-eae2-4db8-a82e-a7e1d1c462e5" (UID: "99df09c3-eae2-4db8-a82e-a7e1d1c462e5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.432060 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-kube-api-access-tkqmk" (OuterVolumeSpecName: "kube-api-access-tkqmk") pod "99df09c3-eae2-4db8-a82e-a7e1d1c462e5" (UID: "99df09c3-eae2-4db8-a82e-a7e1d1c462e5"). InnerVolumeSpecName "kube-api-access-tkqmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.527715 4629 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.527749 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkqmk\" (UniqueName: \"kubernetes.io/projected/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-kube-api-access-tkqmk\") on node \"crc\" DevicePath \"\"" Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.527761 4629 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99df09c3-eae2-4db8-a82e-a7e1d1c462e5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.824883 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ea2b9d9a-81ee-4120-be06-cc8a5684c765","Type":"ContainerStarted","Data":"4218ebfcf035e9bff67694672ccd5f7062873e64f7e3ef55c2ddc6e20506a5f2"} Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.831343 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" event={"ID":"99df09c3-eae2-4db8-a82e-a7e1d1c462e5","Type":"ContainerDied","Data":"4dc746562a886a2e349437170dc4d9a11e6e1bc50ae8b6642429882436139e44"} Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.831390 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4dc746562a886a2e349437170dc4d9a11e6e1bc50ae8b6642429882436139e44" Dec 11 08:39:40 crc kubenswrapper[4629]: I1211 08:39:40.831419 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm" Dec 11 08:39:41 crc kubenswrapper[4629]: I1211 08:39:41.353369 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:41 crc kubenswrapper[4629]: I1211 08:39:41.356955 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-r9x7m" Dec 11 08:39:41 crc kubenswrapper[4629]: I1211 08:39:41.384772 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.384747008 podStartE2EDuration="4.384747008s" podCreationTimestamp="2025-12-11 08:39:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:39:40.848688469 +0000 UTC m=+168.914104117" watchObservedRunningTime="2025-12-11 08:39:41.384747008 +0000 UTC m=+169.450162616" Dec 11 08:39:41 crc kubenswrapper[4629]: I1211 08:39:41.858896 4629 generic.go:334] "Generic (PLEG): container finished" podID="ea2b9d9a-81ee-4120-be06-cc8a5684c765" containerID="4218ebfcf035e9bff67694672ccd5f7062873e64f7e3ef55c2ddc6e20506a5f2" exitCode=0 Dec 11 08:39:41 crc kubenswrapper[4629]: I1211 08:39:41.859000 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ea2b9d9a-81ee-4120-be06-cc8a5684c765","Type":"ContainerDied","Data":"4218ebfcf035e9bff67694672ccd5f7062873e64f7e3ef55c2ddc6e20506a5f2"} Dec 11 08:39:48 crc kubenswrapper[4629]: I1211 08:39:48.534628 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:39:48 crc kubenswrapper[4629]: I1211 08:39:48.856180 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:48 crc kubenswrapper[4629]: I1211 08:39:48.860255 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.101674 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.174664 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea2b9d9a-81ee-4120-be06-cc8a5684c765-kube-api-access\") pod \"ea2b9d9a-81ee-4120-be06-cc8a5684c765\" (UID: \"ea2b9d9a-81ee-4120-be06-cc8a5684c765\") " Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.174787 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea2b9d9a-81ee-4120-be06-cc8a5684c765-kubelet-dir\") pod \"ea2b9d9a-81ee-4120-be06-cc8a5684c765\" (UID: \"ea2b9d9a-81ee-4120-be06-cc8a5684c765\") " Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.175074 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea2b9d9a-81ee-4120-be06-cc8a5684c765-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ea2b9d9a-81ee-4120-be06-cc8a5684c765" (UID: "ea2b9d9a-81ee-4120-be06-cc8a5684c765"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.179560 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea2b9d9a-81ee-4120-be06-cc8a5684c765-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ea2b9d9a-81ee-4120-be06-cc8a5684c765" (UID: "ea2b9d9a-81ee-4120-be06-cc8a5684c765"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.276154 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea2b9d9a-81ee-4120-be06-cc8a5684c765-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.276540 4629 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea2b9d9a-81ee-4120-be06-cc8a5684c765-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.486299 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.486360 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.857276 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.945673 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ea2b9d9a-81ee-4120-be06-cc8a5684c765","Type":"ContainerDied","Data":"0c0040a3e942694d83f783465b1fe67bb874eaa1a62a04f0ef2d74835cbf5c80"} Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.945715 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c0040a3e942694d83f783465b1fe67bb874eaa1a62a04f0ef2d74835cbf5c80" Dec 11 08:39:50 crc kubenswrapper[4629]: I1211 08:39:50.945773 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:39:59 crc kubenswrapper[4629]: I1211 08:39:59.453379 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rlll" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.460268 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 11 08:40:09 crc kubenswrapper[4629]: E1211 08:40:09.462284 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44644f73-1584-404e-a605-01fb14c97921" containerName="pruner" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.462363 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="44644f73-1584-404e-a605-01fb14c97921" containerName="pruner" Dec 11 08:40:09 crc kubenswrapper[4629]: E1211 08:40:09.462442 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99df09c3-eae2-4db8-a82e-a7e1d1c462e5" containerName="collect-profiles" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.462517 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="99df09c3-eae2-4db8-a82e-a7e1d1c462e5" containerName="collect-profiles" Dec 11 08:40:09 crc kubenswrapper[4629]: E1211 08:40:09.462690 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea2b9d9a-81ee-4120-be06-cc8a5684c765" containerName="pruner" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.462763 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea2b9d9a-81ee-4120-be06-cc8a5684c765" containerName="pruner" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.462969 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="99df09c3-eae2-4db8-a82e-a7e1d1c462e5" containerName="collect-profiles" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.463049 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="44644f73-1584-404e-a605-01fb14c97921" containerName="pruner" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.463166 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea2b9d9a-81ee-4120-be06-cc8a5684c765" containerName="pruner" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.463714 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.466390 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.466771 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.468765 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b0de2ef-6a6c-460c-bf25-325cee52f44f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0b0de2ef-6a6c-460c-bf25-325cee52f44f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.468865 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0b0de2ef-6a6c-460c-bf25-325cee52f44f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0b0de2ef-6a6c-460c-bf25-325cee52f44f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.479154 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.570029 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b0de2ef-6a6c-460c-bf25-325cee52f44f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0b0de2ef-6a6c-460c-bf25-325cee52f44f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.570093 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0b0de2ef-6a6c-460c-bf25-325cee52f44f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0b0de2ef-6a6c-460c-bf25-325cee52f44f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.570176 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0b0de2ef-6a6c-460c-bf25-325cee52f44f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0b0de2ef-6a6c-460c-bf25-325cee52f44f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.596267 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b0de2ef-6a6c-460c-bf25-325cee52f44f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0b0de2ef-6a6c-460c-bf25-325cee52f44f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:40:09 crc kubenswrapper[4629]: I1211 08:40:09.783613 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:40:12 crc kubenswrapper[4629]: E1211 08:40:12.027821 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 11 08:40:12 crc kubenswrapper[4629]: E1211 08:40:12.028467 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q47jn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-n2r4l_openshift-marketplace(ccfb1548-f379-4fb3-ba39-b138585900b4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:40:12 crc kubenswrapper[4629]: E1211 08:40:12.030019 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-n2r4l" podUID="ccfb1548-f379-4fb3-ba39-b138585900b4" Dec 11 08:40:12 crc kubenswrapper[4629]: E1211 08:40:12.308223 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-n2r4l" podUID="ccfb1548-f379-4fb3-ba39-b138585900b4" Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.647457 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.649795 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.659640 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.788972 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/695de9c8-1898-48e1-8967-982a5d2cb3ab-kube-api-access\") pod \"installer-9-crc\" (UID: \"695de9c8-1898-48e1-8967-982a5d2cb3ab\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.789447 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/695de9c8-1898-48e1-8967-982a5d2cb3ab-kubelet-dir\") pod \"installer-9-crc\" (UID: \"695de9c8-1898-48e1-8967-982a5d2cb3ab\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.789539 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/695de9c8-1898-48e1-8967-982a5d2cb3ab-var-lock\") pod \"installer-9-crc\" (UID: \"695de9c8-1898-48e1-8967-982a5d2cb3ab\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.890367 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/695de9c8-1898-48e1-8967-982a5d2cb3ab-kube-api-access\") pod \"installer-9-crc\" (UID: \"695de9c8-1898-48e1-8967-982a5d2cb3ab\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.890425 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/695de9c8-1898-48e1-8967-982a5d2cb3ab-kubelet-dir\") pod \"installer-9-crc\" (UID: \"695de9c8-1898-48e1-8967-982a5d2cb3ab\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.890485 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/695de9c8-1898-48e1-8967-982a5d2cb3ab-var-lock\") pod \"installer-9-crc\" (UID: \"695de9c8-1898-48e1-8967-982a5d2cb3ab\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.890591 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/695de9c8-1898-48e1-8967-982a5d2cb3ab-var-lock\") pod \"installer-9-crc\" (UID: \"695de9c8-1898-48e1-8967-982a5d2cb3ab\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.890635 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/695de9c8-1898-48e1-8967-982a5d2cb3ab-kubelet-dir\") pod \"installer-9-crc\" (UID: \"695de9c8-1898-48e1-8967-982a5d2cb3ab\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.909389 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/695de9c8-1898-48e1-8967-982a5d2cb3ab-kube-api-access\") pod \"installer-9-crc\" (UID: \"695de9c8-1898-48e1-8967-982a5d2cb3ab\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:40:13 crc kubenswrapper[4629]: I1211 08:40:13.992033 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:40:17 crc kubenswrapper[4629]: E1211 08:40:17.324640 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 11 08:40:17 crc kubenswrapper[4629]: E1211 08:40:17.325539 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wwdfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-69vgp_openshift-marketplace(ecaf1148-6695-4cdf-9ffc-06cab2716241): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:40:17 crc kubenswrapper[4629]: E1211 08:40:17.326963 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-69vgp" podUID="ecaf1148-6695-4cdf-9ffc-06cab2716241" Dec 11 08:40:17 crc kubenswrapper[4629]: E1211 08:40:17.426511 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 11 08:40:17 crc kubenswrapper[4629]: E1211 08:40:17.426639 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c8q4g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-kzp9g_openshift-marketplace(7adee9e0-edb2-4ea3-afa6-d5759440a79a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:40:17 crc kubenswrapper[4629]: E1211 08:40:17.428875 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-kzp9g" podUID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" Dec 11 08:40:19 crc kubenswrapper[4629]: E1211 08:40:19.610890 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-69vgp" podUID="ecaf1148-6695-4cdf-9ffc-06cab2716241" Dec 11 08:40:19 crc kubenswrapper[4629]: E1211 08:40:19.611112 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-kzp9g" podUID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" Dec 11 08:40:19 crc kubenswrapper[4629]: E1211 08:40:19.714199 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 11 08:40:19 crc kubenswrapper[4629]: E1211 08:40:19.714348 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tm8t6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-8f998_openshift-marketplace(c60b1455-35ba-47a0-a4d8-016c651e8c22): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:40:19 crc kubenswrapper[4629]: E1211 08:40:19.716690 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-8f998" podUID="c60b1455-35ba-47a0-a4d8-016c651e8c22" Dec 11 08:40:19 crc kubenswrapper[4629]: E1211 08:40:19.745004 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 11 08:40:19 crc kubenswrapper[4629]: E1211 08:40:19.745140 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xwjpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-587qx_openshift-marketplace(1cc54f96-ce8c-4efe-9c1f-ac612c9ab366): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:40:19 crc kubenswrapper[4629]: E1211 08:40:19.746956 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-587qx" podUID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" Dec 11 08:40:19 crc kubenswrapper[4629]: E1211 08:40:19.795503 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 11 08:40:19 crc kubenswrapper[4629]: E1211 08:40:19.795661 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v4qzq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-c6kzc_openshift-marketplace(69c02574-1a57-4d6c-aa9c-8a85c6acd80f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:40:19 crc kubenswrapper[4629]: E1211 08:40:19.796860 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-c6kzc" podUID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" Dec 11 08:40:20 crc kubenswrapper[4629]: I1211 08:40:20.485968 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:40:20 crc kubenswrapper[4629]: I1211 08:40:20.486681 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:40:20 crc kubenswrapper[4629]: I1211 08:40:20.486749 4629 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:40:20 crc kubenswrapper[4629]: I1211 08:40:20.488302 4629 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7"} pod="openshift-machine-config-operator/machine-config-daemon-wbld9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:40:20 crc kubenswrapper[4629]: I1211 08:40:20.488451 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" containerID="cri-o://df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7" gracePeriod=600 Dec 11 08:40:21 crc kubenswrapper[4629]: I1211 08:40:21.447726 4629 generic.go:334] "Generic (PLEG): container finished" podID="847b50b9-0093-404f-b0dd-63c814d8986b" containerID="df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7" exitCode=0 Dec 11 08:40:21 crc kubenswrapper[4629]: I1211 08:40:21.447768 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerDied","Data":"df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7"} Dec 11 08:40:24 crc kubenswrapper[4629]: E1211 08:40:24.870241 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-c6kzc" podUID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" Dec 11 08:40:24 crc kubenswrapper[4629]: E1211 08:40:24.878559 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-8f998" podUID="c60b1455-35ba-47a0-a4d8-016c651e8c22" Dec 11 08:40:24 crc kubenswrapper[4629]: E1211 08:40:24.878775 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-587qx" podUID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" Dec 11 08:40:27 crc kubenswrapper[4629]: E1211 08:40:27.324115 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 11 08:40:27 crc kubenswrapper[4629]: E1211 08:40:27.324543 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4kbnb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-q8vkr_openshift-marketplace(28a0959a-1692-4c2e-83c3-54627cd1d53a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:40:27 crc kubenswrapper[4629]: E1211 08:40:27.326547 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-q8vkr" podUID="28a0959a-1692-4c2e-83c3-54627cd1d53a" Dec 11 08:40:27 crc kubenswrapper[4629]: I1211 08:40:27.398925 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 11 08:40:27 crc kubenswrapper[4629]: E1211 08:40:27.406434 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 11 08:40:27 crc kubenswrapper[4629]: E1211 08:40:27.406591 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6zvjc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bwgwp_openshift-marketplace(1451fb38-b06c-4b5c-bb43-8b571b8e3302): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:40:27 crc kubenswrapper[4629]: E1211 08:40:27.407919 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bwgwp" podUID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" Dec 11 08:40:27 crc kubenswrapper[4629]: I1211 08:40:27.498418 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 11 08:40:27 crc kubenswrapper[4629]: I1211 08:40:27.500797 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0b0de2ef-6a6c-460c-bf25-325cee52f44f","Type":"ContainerStarted","Data":"7cf69350cf6a1183bb9e6cac98832cbdb4bde608a697ca9db9b2e86405ead34d"} Dec 11 08:40:27 crc kubenswrapper[4629]: I1211 08:40:27.511144 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"8ba210253304f5487d01920ea72fb77a38de3a8bfc04263290603cecc0511379"} Dec 11 08:40:27 crc kubenswrapper[4629]: E1211 08:40:27.512874 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-q8vkr" podUID="28a0959a-1692-4c2e-83c3-54627cd1d53a" Dec 11 08:40:27 crc kubenswrapper[4629]: E1211 08:40:27.513462 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bwgwp" podUID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" Dec 11 08:40:28 crc kubenswrapper[4629]: I1211 08:40:28.517910 4629 generic.go:334] "Generic (PLEG): container finished" podID="0b0de2ef-6a6c-460c-bf25-325cee52f44f" containerID="f381253808966dd4f64276851a8ee3b384c77a986d28fa3f7325e2a62229e619" exitCode=0 Dec 11 08:40:28 crc kubenswrapper[4629]: I1211 08:40:28.518952 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0b0de2ef-6a6c-460c-bf25-325cee52f44f","Type":"ContainerDied","Data":"f381253808966dd4f64276851a8ee3b384c77a986d28fa3f7325e2a62229e619"} Dec 11 08:40:28 crc kubenswrapper[4629]: I1211 08:40:28.521276 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"695de9c8-1898-48e1-8967-982a5d2cb3ab","Type":"ContainerStarted","Data":"fe20e691db6497318b74c87a3fd4b77dcd02b5e8f999b7624f94e861f7ec1e74"} Dec 11 08:40:28 crc kubenswrapper[4629]: I1211 08:40:28.521442 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"695de9c8-1898-48e1-8967-982a5d2cb3ab","Type":"ContainerStarted","Data":"eef351cd94a652798e5e93e09b019efc3cdb21d0b2a857a797d57ce7b322ef98"} Dec 11 08:40:28 crc kubenswrapper[4629]: I1211 08:40:28.549524 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=15.549491079 podStartE2EDuration="15.549491079s" podCreationTimestamp="2025-12-11 08:40:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:40:28.54888303 +0000 UTC m=+216.614298638" watchObservedRunningTime="2025-12-11 08:40:28.549491079 +0000 UTC m=+216.614906687" Dec 11 08:40:29 crc kubenswrapper[4629]: I1211 08:40:29.533920 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2r4l" event={"ID":"ccfb1548-f379-4fb3-ba39-b138585900b4","Type":"ContainerStarted","Data":"373d7b045bc8fae5fb716e3f138c163ef5b68f2199a9952045f1fb242c1aa0c7"} Dec 11 08:40:29 crc kubenswrapper[4629]: I1211 08:40:29.796890 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:40:29 crc kubenswrapper[4629]: I1211 08:40:29.912529 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0b0de2ef-6a6c-460c-bf25-325cee52f44f-kubelet-dir\") pod \"0b0de2ef-6a6c-460c-bf25-325cee52f44f\" (UID: \"0b0de2ef-6a6c-460c-bf25-325cee52f44f\") " Dec 11 08:40:29 crc kubenswrapper[4629]: I1211 08:40:29.912594 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b0de2ef-6a6c-460c-bf25-325cee52f44f-kube-api-access\") pod \"0b0de2ef-6a6c-460c-bf25-325cee52f44f\" (UID: \"0b0de2ef-6a6c-460c-bf25-325cee52f44f\") " Dec 11 08:40:29 crc kubenswrapper[4629]: I1211 08:40:29.912683 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0b0de2ef-6a6c-460c-bf25-325cee52f44f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0b0de2ef-6a6c-460c-bf25-325cee52f44f" (UID: "0b0de2ef-6a6c-460c-bf25-325cee52f44f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:40:29 crc kubenswrapper[4629]: I1211 08:40:29.913049 4629 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0b0de2ef-6a6c-460c-bf25-325cee52f44f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:29 crc kubenswrapper[4629]: I1211 08:40:29.918415 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b0de2ef-6a6c-460c-bf25-325cee52f44f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b0de2ef-6a6c-460c-bf25-325cee52f44f" (UID: "0b0de2ef-6a6c-460c-bf25-325cee52f44f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:40:30 crc kubenswrapper[4629]: I1211 08:40:30.014562 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b0de2ef-6a6c-460c-bf25-325cee52f44f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:30 crc kubenswrapper[4629]: I1211 08:40:30.551603 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0b0de2ef-6a6c-460c-bf25-325cee52f44f","Type":"ContainerDied","Data":"7cf69350cf6a1183bb9e6cac98832cbdb4bde608a697ca9db9b2e86405ead34d"} Dec 11 08:40:30 crc kubenswrapper[4629]: I1211 08:40:30.551643 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cf69350cf6a1183bb9e6cac98832cbdb4bde608a697ca9db9b2e86405ead34d" Dec 11 08:40:30 crc kubenswrapper[4629]: I1211 08:40:30.551651 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:40:30 crc kubenswrapper[4629]: I1211 08:40:30.557018 4629 generic.go:334] "Generic (PLEG): container finished" podID="ccfb1548-f379-4fb3-ba39-b138585900b4" containerID="373d7b045bc8fae5fb716e3f138c163ef5b68f2199a9952045f1fb242c1aa0c7" exitCode=0 Dec 11 08:40:30 crc kubenswrapper[4629]: I1211 08:40:30.557044 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2r4l" event={"ID":"ccfb1548-f379-4fb3-ba39-b138585900b4","Type":"ContainerDied","Data":"373d7b045bc8fae5fb716e3f138c163ef5b68f2199a9952045f1fb242c1aa0c7"} Dec 11 08:40:32 crc kubenswrapper[4629]: I1211 08:40:32.568079 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69vgp" event={"ID":"ecaf1148-6695-4cdf-9ffc-06cab2716241","Type":"ContainerStarted","Data":"f3b6f6792b49d9d807f7359fbfb32c85663c623b9032a863acb3a00aef93be47"} Dec 11 08:40:32 crc kubenswrapper[4629]: I1211 08:40:32.570413 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2r4l" event={"ID":"ccfb1548-f379-4fb3-ba39-b138585900b4","Type":"ContainerStarted","Data":"0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa"} Dec 11 08:40:32 crc kubenswrapper[4629]: I1211 08:40:32.607152 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n2r4l" podStartSLOduration=3.478841735 podStartE2EDuration="1m4.607129714s" podCreationTimestamp="2025-12-11 08:39:28 +0000 UTC" firstStartedPulling="2025-12-11 08:39:30.423851033 +0000 UTC m=+158.489266641" lastFinishedPulling="2025-12-11 08:40:31.552139012 +0000 UTC m=+219.617554620" observedRunningTime="2025-12-11 08:40:32.606504945 +0000 UTC m=+220.671920553" watchObservedRunningTime="2025-12-11 08:40:32.607129714 +0000 UTC m=+220.672545322" Dec 11 08:40:33 crc kubenswrapper[4629]: I1211 08:40:33.577040 4629 generic.go:334] "Generic (PLEG): container finished" podID="ecaf1148-6695-4cdf-9ffc-06cab2716241" containerID="f3b6f6792b49d9d807f7359fbfb32c85663c623b9032a863acb3a00aef93be47" exitCode=0 Dec 11 08:40:33 crc kubenswrapper[4629]: I1211 08:40:33.577081 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69vgp" event={"ID":"ecaf1148-6695-4cdf-9ffc-06cab2716241","Type":"ContainerDied","Data":"f3b6f6792b49d9d807f7359fbfb32c85663c623b9032a863acb3a00aef93be47"} Dec 11 08:40:38 crc kubenswrapper[4629]: I1211 08:40:38.692667 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:40:38 crc kubenswrapper[4629]: I1211 08:40:38.693311 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:40:38 crc kubenswrapper[4629]: I1211 08:40:38.759540 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:40:39 crc kubenswrapper[4629]: I1211 08:40:39.664014 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:40:41 crc kubenswrapper[4629]: I1211 08:40:41.624488 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69vgp" event={"ID":"ecaf1148-6695-4cdf-9ffc-06cab2716241","Type":"ContainerStarted","Data":"59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae"} Dec 11 08:40:41 crc kubenswrapper[4629]: I1211 08:40:41.626060 4629 generic.go:334] "Generic (PLEG): container finished" podID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" containerID="573ae2b78f910b9947cb62489cac6be72a938d8011058c8a70e11d3a2f8dc93c" exitCode=0 Dec 11 08:40:41 crc kubenswrapper[4629]: I1211 08:40:41.626115 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzp9g" event={"ID":"7adee9e0-edb2-4ea3-afa6-d5759440a79a","Type":"ContainerDied","Data":"573ae2b78f910b9947cb62489cac6be72a938d8011058c8a70e11d3a2f8dc93c"} Dec 11 08:40:41 crc kubenswrapper[4629]: I1211 08:40:41.644003 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-69vgp" podStartSLOduration=4.6805369169999995 podStartE2EDuration="1m10.643972565s" podCreationTimestamp="2025-12-11 08:39:31 +0000 UTC" firstStartedPulling="2025-12-11 08:39:33.52758708 +0000 UTC m=+161.593002688" lastFinishedPulling="2025-12-11 08:40:39.491022728 +0000 UTC m=+227.556438336" observedRunningTime="2025-12-11 08:40:41.640801156 +0000 UTC m=+229.706216774" watchObservedRunningTime="2025-12-11 08:40:41.643972565 +0000 UTC m=+229.709388173" Dec 11 08:40:41 crc kubenswrapper[4629]: I1211 08:40:41.837151 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:40:41 crc kubenswrapper[4629]: I1211 08:40:41.838396 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:40:42 crc kubenswrapper[4629]: I1211 08:40:42.633401 4629 generic.go:334] "Generic (PLEG): container finished" podID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" containerID="eaa8a0b60e5d7c01de30a63bd007e68ba158c2377c6ec688a89092e0db0c798f" exitCode=0 Dec 11 08:40:42 crc kubenswrapper[4629]: I1211 08:40:42.633479 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-587qx" event={"ID":"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366","Type":"ContainerDied","Data":"eaa8a0b60e5d7c01de30a63bd007e68ba158c2377c6ec688a89092e0db0c798f"} Dec 11 08:40:42 crc kubenswrapper[4629]: I1211 08:40:42.638778 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzp9g" event={"ID":"7adee9e0-edb2-4ea3-afa6-d5759440a79a","Type":"ContainerStarted","Data":"4c11a89f86ca71fe2519d7bf15ac54bb7bb2b3fb1b7adaef45558434cfb0d2bc"} Dec 11 08:40:42 crc kubenswrapper[4629]: I1211 08:40:42.646577 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwgwp" event={"ID":"1451fb38-b06c-4b5c-bb43-8b571b8e3302","Type":"ContainerStarted","Data":"9f62ec2ab1a5d1dcdf6a3d4c53b894f9ef9306009cc6215146d8351366b9b42a"} Dec 11 08:40:42 crc kubenswrapper[4629]: I1211 08:40:42.651163 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8f998" event={"ID":"c60b1455-35ba-47a0-a4d8-016c651e8c22","Type":"ContainerStarted","Data":"f29b61cfc37fa66bffae8b0f9908855b925c5a6acb348820cb704096cd53daed"} Dec 11 08:40:42 crc kubenswrapper[4629]: I1211 08:40:42.725569 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kzp9g" podStartSLOduration=3.005899763 podStartE2EDuration="1m14.725553277s" podCreationTimestamp="2025-12-11 08:39:28 +0000 UTC" firstStartedPulling="2025-12-11 08:39:30.423480572 +0000 UTC m=+158.488896180" lastFinishedPulling="2025-12-11 08:40:42.143134086 +0000 UTC m=+230.208549694" observedRunningTime="2025-12-11 08:40:42.704336834 +0000 UTC m=+230.769752462" watchObservedRunningTime="2025-12-11 08:40:42.725553277 +0000 UTC m=+230.790968885" Dec 11 08:40:42 crc kubenswrapper[4629]: I1211 08:40:42.884436 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-69vgp" podUID="ecaf1148-6695-4cdf-9ffc-06cab2716241" containerName="registry-server" probeResult="failure" output=< Dec 11 08:40:42 crc kubenswrapper[4629]: timeout: failed to connect service ":50051" within 1s Dec 11 08:40:42 crc kubenswrapper[4629]: > Dec 11 08:40:43 crc kubenswrapper[4629]: I1211 08:40:43.907801 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8vkr" event={"ID":"28a0959a-1692-4c2e-83c3-54627cd1d53a","Type":"ContainerStarted","Data":"e657d64611c3d908b9654ed4e7933bcef9fe15a08cc162f71077d398bb64d400"} Dec 11 08:40:43 crc kubenswrapper[4629]: I1211 08:40:43.912637 4629 generic.go:334] "Generic (PLEG): container finished" podID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" containerID="9f62ec2ab1a5d1dcdf6a3d4c53b894f9ef9306009cc6215146d8351366b9b42a" exitCode=0 Dec 11 08:40:43 crc kubenswrapper[4629]: I1211 08:40:43.912723 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwgwp" event={"ID":"1451fb38-b06c-4b5c-bb43-8b571b8e3302","Type":"ContainerDied","Data":"9f62ec2ab1a5d1dcdf6a3d4c53b894f9ef9306009cc6215146d8351366b9b42a"} Dec 11 08:40:43 crc kubenswrapper[4629]: I1211 08:40:43.915585 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8f998" event={"ID":"c60b1455-35ba-47a0-a4d8-016c651e8c22","Type":"ContainerDied","Data":"f29b61cfc37fa66bffae8b0f9908855b925c5a6acb348820cb704096cd53daed"} Dec 11 08:40:43 crc kubenswrapper[4629]: I1211 08:40:43.915501 4629 generic.go:334] "Generic (PLEG): container finished" podID="c60b1455-35ba-47a0-a4d8-016c651e8c22" containerID="f29b61cfc37fa66bffae8b0f9908855b925c5a6acb348820cb704096cd53daed" exitCode=0 Dec 11 08:40:43 crc kubenswrapper[4629]: I1211 08:40:43.919074 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kzc" event={"ID":"69c02574-1a57-4d6c-aa9c-8a85c6acd80f","Type":"ContainerStarted","Data":"f4c1b5f7a2bc031a2d485d62b86b14fc29a06078466c8704b52d1776194588bd"} Dec 11 08:40:44 crc kubenswrapper[4629]: I1211 08:40:44.928971 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8f998" event={"ID":"c60b1455-35ba-47a0-a4d8-016c651e8c22","Type":"ContainerStarted","Data":"38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f"} Dec 11 08:40:44 crc kubenswrapper[4629]: I1211 08:40:44.931146 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-587qx" event={"ID":"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366","Type":"ContainerStarted","Data":"0582e7740e27b156bf57bd34ad3a530d25892a96a9c5d1ce57edfa65237550c0"} Dec 11 08:40:44 crc kubenswrapper[4629]: I1211 08:40:44.933705 4629 generic.go:334] "Generic (PLEG): container finished" podID="28a0959a-1692-4c2e-83c3-54627cd1d53a" containerID="e657d64611c3d908b9654ed4e7933bcef9fe15a08cc162f71077d398bb64d400" exitCode=0 Dec 11 08:40:44 crc kubenswrapper[4629]: I1211 08:40:44.933765 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8vkr" event={"ID":"28a0959a-1692-4c2e-83c3-54627cd1d53a","Type":"ContainerDied","Data":"e657d64611c3d908b9654ed4e7933bcef9fe15a08cc162f71077d398bb64d400"} Dec 11 08:40:44 crc kubenswrapper[4629]: I1211 08:40:44.941260 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwgwp" event={"ID":"1451fb38-b06c-4b5c-bb43-8b571b8e3302","Type":"ContainerStarted","Data":"33976ac375076b4955816ac01d2d04dac21238ee50ee5fed8940b8aeee10e8df"} Dec 11 08:40:44 crc kubenswrapper[4629]: I1211 08:40:44.988196 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8f998" podStartSLOduration=3.816471554 podStartE2EDuration="1m16.988175612s" podCreationTimestamp="2025-12-11 08:39:28 +0000 UTC" firstStartedPulling="2025-12-11 08:39:31.431044488 +0000 UTC m=+159.496460096" lastFinishedPulling="2025-12-11 08:40:44.602748546 +0000 UTC m=+232.668164154" observedRunningTime="2025-12-11 08:40:44.961265031 +0000 UTC m=+233.026680639" watchObservedRunningTime="2025-12-11 08:40:44.988175612 +0000 UTC m=+233.053591210" Dec 11 08:40:44 crc kubenswrapper[4629]: I1211 08:40:44.989726 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bwgwp" podStartSLOduration=3.2568140789999998 podStartE2EDuration="1m13.989717291s" podCreationTimestamp="2025-12-11 08:39:31 +0000 UTC" firstStartedPulling="2025-12-11 08:39:33.56668301 +0000 UTC m=+161.632098618" lastFinishedPulling="2025-12-11 08:40:44.299586222 +0000 UTC m=+232.365001830" observedRunningTime="2025-12-11 08:40:44.987288345 +0000 UTC m=+233.052703953" watchObservedRunningTime="2025-12-11 08:40:44.989717291 +0000 UTC m=+233.055132899" Dec 11 08:40:45 crc kubenswrapper[4629]: I1211 08:40:45.042088 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-587qx" podStartSLOduration=4.930154044 podStartE2EDuration="1m17.042067817s" podCreationTimestamp="2025-12-11 08:39:28 +0000 UTC" firstStartedPulling="2025-12-11 08:39:31.450875151 +0000 UTC m=+159.516290759" lastFinishedPulling="2025-12-11 08:40:43.562788924 +0000 UTC m=+231.628204532" observedRunningTime="2025-12-11 08:40:45.040044324 +0000 UTC m=+233.105459932" watchObservedRunningTime="2025-12-11 08:40:45.042067817 +0000 UTC m=+233.107483425" Dec 11 08:40:45 crc kubenswrapper[4629]: I1211 08:40:45.949650 4629 generic.go:334] "Generic (PLEG): container finished" podID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" containerID="f4c1b5f7a2bc031a2d485d62b86b14fc29a06078466c8704b52d1776194588bd" exitCode=0 Dec 11 08:40:45 crc kubenswrapper[4629]: I1211 08:40:45.949782 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kzc" event={"ID":"69c02574-1a57-4d6c-aa9c-8a85c6acd80f","Type":"ContainerDied","Data":"f4c1b5f7a2bc031a2d485d62b86b14fc29a06078466c8704b52d1776194588bd"} Dec 11 08:40:45 crc kubenswrapper[4629]: I1211 08:40:45.954103 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8vkr" event={"ID":"28a0959a-1692-4c2e-83c3-54627cd1d53a","Type":"ContainerStarted","Data":"bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420"} Dec 11 08:40:45 crc kubenswrapper[4629]: I1211 08:40:45.999346 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q8vkr" podStartSLOduration=2.993743191 podStartE2EDuration="1m15.999325184s" podCreationTimestamp="2025-12-11 08:39:30 +0000 UTC" firstStartedPulling="2025-12-11 08:39:32.505529232 +0000 UTC m=+160.570944840" lastFinishedPulling="2025-12-11 08:40:45.511111225 +0000 UTC m=+233.576526833" observedRunningTime="2025-12-11 08:40:45.995797513 +0000 UTC m=+234.061213121" watchObservedRunningTime="2025-12-11 08:40:45.999325184 +0000 UTC m=+234.064740792" Dec 11 08:40:47 crc kubenswrapper[4629]: I1211 08:40:47.979197 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kzc" event={"ID":"69c02574-1a57-4d6c-aa9c-8a85c6acd80f","Type":"ContainerStarted","Data":"2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c"} Dec 11 08:40:49 crc kubenswrapper[4629]: I1211 08:40:49.091021 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:40:49 crc kubenswrapper[4629]: I1211 08:40:49.091070 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:40:49 crc kubenswrapper[4629]: I1211 08:40:49.129661 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:40:49 crc kubenswrapper[4629]: I1211 08:40:49.134734 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8f998" Dec 11 08:40:49 crc kubenswrapper[4629]: I1211 08:40:49.134821 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8f998" Dec 11 08:40:49 crc kubenswrapper[4629]: I1211 08:40:49.159148 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c6kzc" podStartSLOduration=4.784164378 podStartE2EDuration="1m18.159129929s" podCreationTimestamp="2025-12-11 08:39:31 +0000 UTC" firstStartedPulling="2025-12-11 08:39:33.559357095 +0000 UTC m=+161.624772703" lastFinishedPulling="2025-12-11 08:40:46.934322646 +0000 UTC m=+234.999738254" observedRunningTime="2025-12-11 08:40:48.114195552 +0000 UTC m=+236.179611180" watchObservedRunningTime="2025-12-11 08:40:49.159129929 +0000 UTC m=+237.224545537" Dec 11 08:40:49 crc kubenswrapper[4629]: I1211 08:40:49.174440 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8f998" Dec 11 08:40:49 crc kubenswrapper[4629]: I1211 08:40:49.249985 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-587qx" Dec 11 08:40:49 crc kubenswrapper[4629]: I1211 08:40:49.250234 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-587qx" Dec 11 08:40:49 crc kubenswrapper[4629]: I1211 08:40:49.293947 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-587qx" Dec 11 08:40:50 crc kubenswrapper[4629]: I1211 08:40:50.035073 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8f998" Dec 11 08:40:50 crc kubenswrapper[4629]: I1211 08:40:50.039958 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:40:50 crc kubenswrapper[4629]: I1211 08:40:50.047102 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-587qx" Dec 11 08:40:51 crc kubenswrapper[4629]: I1211 08:40:51.042275 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:40:51 crc kubenswrapper[4629]: I1211 08:40:51.042323 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:40:51 crc kubenswrapper[4629]: I1211 08:40:51.088275 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:40:51 crc kubenswrapper[4629]: I1211 08:40:51.459645 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:40:51 crc kubenswrapper[4629]: I1211 08:40:51.459716 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:40:51 crc kubenswrapper[4629]: I1211 08:40:51.497871 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:40:51 crc kubenswrapper[4629]: I1211 08:40:51.884835 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:40:51 crc kubenswrapper[4629]: I1211 08:40:51.943992 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kzp9g"] Dec 11 08:40:51 crc kubenswrapper[4629]: I1211 08:40:51.949636 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:40:52 crc kubenswrapper[4629]: I1211 08:40:52.002046 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kzp9g" podUID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" containerName="registry-server" containerID="cri-o://4c11a89f86ca71fe2519d7bf15ac54bb7bb2b3fb1b7adaef45558434cfb0d2bc" gracePeriod=2 Dec 11 08:40:52 crc kubenswrapper[4629]: I1211 08:40:52.060070 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:40:52 crc kubenswrapper[4629]: I1211 08:40:52.110762 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:40:52 crc kubenswrapper[4629]: I1211 08:40:52.110829 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:40:52 crc kubenswrapper[4629]: I1211 08:40:52.617133 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:40:53 crc kubenswrapper[4629]: I1211 08:40:53.011367 4629 generic.go:334] "Generic (PLEG): container finished" podID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" containerID="4c11a89f86ca71fe2519d7bf15ac54bb7bb2b3fb1b7adaef45558434cfb0d2bc" exitCode=0 Dec 11 08:40:53 crc kubenswrapper[4629]: I1211 08:40:53.011744 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzp9g" event={"ID":"7adee9e0-edb2-4ea3-afa6-d5759440a79a","Type":"ContainerDied","Data":"4c11a89f86ca71fe2519d7bf15ac54bb7bb2b3fb1b7adaef45558434cfb0d2bc"} Dec 11 08:40:53 crc kubenswrapper[4629]: I1211 08:40:53.153489 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c6kzc" podUID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" containerName="registry-server" probeResult="failure" output=< Dec 11 08:40:53 crc kubenswrapper[4629]: timeout: failed to connect service ":50051" within 1s Dec 11 08:40:53 crc kubenswrapper[4629]: > Dec 11 08:40:53 crc kubenswrapper[4629]: I1211 08:40:53.738534 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-587qx"] Dec 11 08:40:53 crc kubenswrapper[4629]: I1211 08:40:53.738812 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-587qx" podUID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" containerName="registry-server" containerID="cri-o://0582e7740e27b156bf57bd34ad3a530d25892a96a9c5d1ce57edfa65237550c0" gracePeriod=2 Dec 11 08:40:54 crc kubenswrapper[4629]: I1211 08:40:54.298385 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:40:54 crc kubenswrapper[4629]: I1211 08:40:54.342805 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwgwp"] Dec 11 08:40:54 crc kubenswrapper[4629]: I1211 08:40:54.343100 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bwgwp" podUID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" containerName="registry-server" containerID="cri-o://33976ac375076b4955816ac01d2d04dac21238ee50ee5fed8940b8aeee10e8df" gracePeriod=2 Dec 11 08:40:54 crc kubenswrapper[4629]: I1211 08:40:54.455763 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7adee9e0-edb2-4ea3-afa6-d5759440a79a-utilities\") pod \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\" (UID: \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\") " Dec 11 08:40:54 crc kubenswrapper[4629]: I1211 08:40:54.457410 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7adee9e0-edb2-4ea3-afa6-d5759440a79a-catalog-content\") pod \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\" (UID: \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\") " Dec 11 08:40:54 crc kubenswrapper[4629]: I1211 08:40:54.462972 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8q4g\" (UniqueName: \"kubernetes.io/projected/7adee9e0-edb2-4ea3-afa6-d5759440a79a-kube-api-access-c8q4g\") pod \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\" (UID: \"7adee9e0-edb2-4ea3-afa6-d5759440a79a\") " Dec 11 08:40:54 crc kubenswrapper[4629]: I1211 08:40:54.457333 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7adee9e0-edb2-4ea3-afa6-d5759440a79a-utilities" (OuterVolumeSpecName: "utilities") pod "7adee9e0-edb2-4ea3-afa6-d5759440a79a" (UID: "7adee9e0-edb2-4ea3-afa6-d5759440a79a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:40:54 crc kubenswrapper[4629]: I1211 08:40:54.463654 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7adee9e0-edb2-4ea3-afa6-d5759440a79a-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:54 crc kubenswrapper[4629]: I1211 08:40:54.468421 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7adee9e0-edb2-4ea3-afa6-d5759440a79a-kube-api-access-c8q4g" (OuterVolumeSpecName: "kube-api-access-c8q4g") pod "7adee9e0-edb2-4ea3-afa6-d5759440a79a" (UID: "7adee9e0-edb2-4ea3-afa6-d5759440a79a"). InnerVolumeSpecName "kube-api-access-c8q4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:40:54 crc kubenswrapper[4629]: I1211 08:40:54.516773 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7adee9e0-edb2-4ea3-afa6-d5759440a79a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7adee9e0-edb2-4ea3-afa6-d5759440a79a" (UID: "7adee9e0-edb2-4ea3-afa6-d5759440a79a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:40:54 crc kubenswrapper[4629]: I1211 08:40:54.564418 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7adee9e0-edb2-4ea3-afa6-d5759440a79a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:54 crc kubenswrapper[4629]: I1211 08:40:54.564463 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8q4g\" (UniqueName: \"kubernetes.io/projected/7adee9e0-edb2-4ea3-afa6-d5759440a79a-kube-api-access-c8q4g\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.027397 4629 generic.go:334] "Generic (PLEG): container finished" podID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" containerID="0582e7740e27b156bf57bd34ad3a530d25892a96a9c5d1ce57edfa65237550c0" exitCode=0 Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.027460 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-587qx" event={"ID":"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366","Type":"ContainerDied","Data":"0582e7740e27b156bf57bd34ad3a530d25892a96a9c5d1ce57edfa65237550c0"} Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.030166 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzp9g" event={"ID":"7adee9e0-edb2-4ea3-afa6-d5759440a79a","Type":"ContainerDied","Data":"cc5fd0e5bb95ee30f7e59883c2192b5bf241ec02a35d75a0633097f59fbad50f"} Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.030425 4629 scope.go:117] "RemoveContainer" containerID="4c11a89f86ca71fe2519d7bf15ac54bb7bb2b3fb1b7adaef45558434cfb0d2bc" Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.030241 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzp9g" Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.060383 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kzp9g"] Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.066971 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kzp9g"] Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.162569 4629 scope.go:117] "RemoveContainer" containerID="573ae2b78f910b9947cb62489cac6be72a938d8011058c8a70e11d3a2f8dc93c" Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.178334 4629 scope.go:117] "RemoveContainer" containerID="c8393d9659441fe47b8496ca347c8a4618a1a881d2a3641de0be1a416f8943ff" Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.637417 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-587qx" Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.777428 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-utilities\") pod \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\" (UID: \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\") " Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.778760 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-catalog-content\") pod \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\" (UID: \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\") " Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.778803 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwjpv\" (UniqueName: \"kubernetes.io/projected/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-kube-api-access-xwjpv\") pod \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\" (UID: \"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366\") " Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.778690 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-utilities" (OuterVolumeSpecName: "utilities") pod "1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" (UID: "1cc54f96-ce8c-4efe-9c1f-ac612c9ab366"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.784027 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-kube-api-access-xwjpv" (OuterVolumeSpecName: "kube-api-access-xwjpv") pod "1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" (UID: "1cc54f96-ce8c-4efe-9c1f-ac612c9ab366"). InnerVolumeSpecName "kube-api-access-xwjpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.826229 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" (UID: "1cc54f96-ce8c-4efe-9c1f-ac612c9ab366"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.889607 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.889657 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:55 crc kubenswrapper[4629]: I1211 08:40:55.889670 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwjpv\" (UniqueName: \"kubernetes.io/projected/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366-kube-api-access-xwjpv\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.037414 4629 generic.go:334] "Generic (PLEG): container finished" podID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" containerID="33976ac375076b4955816ac01d2d04dac21238ee50ee5fed8940b8aeee10e8df" exitCode=0 Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.037473 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwgwp" event={"ID":"1451fb38-b06c-4b5c-bb43-8b571b8e3302","Type":"ContainerDied","Data":"33976ac375076b4955816ac01d2d04dac21238ee50ee5fed8940b8aeee10e8df"} Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.044071 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-587qx" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.044106 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-587qx" event={"ID":"1cc54f96-ce8c-4efe-9c1f-ac612c9ab366","Type":"ContainerDied","Data":"7595dae8a7dd0b1dd85c49ec4a4aa379c86901871e38b9f5c62b568f369955e1"} Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.044192 4629 scope.go:117] "RemoveContainer" containerID="0582e7740e27b156bf57bd34ad3a530d25892a96a9c5d1ce57edfa65237550c0" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.064165 4629 scope.go:117] "RemoveContainer" containerID="eaa8a0b60e5d7c01de30a63bd007e68ba158c2377c6ec688a89092e0db0c798f" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.078824 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-587qx"] Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.081629 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-587qx"] Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.104068 4629 scope.go:117] "RemoveContainer" containerID="550b17b57296d5ebe8356dc21396dcd20fa5be9c4daff2e7600edb9886fe783f" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.214461 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" path="/var/lib/kubelet/pods/1cc54f96-ce8c-4efe-9c1f-ac612c9ab366/volumes" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.215430 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" path="/var/lib/kubelet/pods/7adee9e0-edb2-4ea3-afa6-d5759440a79a/volumes" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.312667 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.399024 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1451fb38-b06c-4b5c-bb43-8b571b8e3302-catalog-content\") pod \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\" (UID: \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\") " Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.399116 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zvjc\" (UniqueName: \"kubernetes.io/projected/1451fb38-b06c-4b5c-bb43-8b571b8e3302-kube-api-access-6zvjc\") pod \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\" (UID: \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\") " Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.399187 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1451fb38-b06c-4b5c-bb43-8b571b8e3302-utilities\") pod \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\" (UID: \"1451fb38-b06c-4b5c-bb43-8b571b8e3302\") " Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.400035 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1451fb38-b06c-4b5c-bb43-8b571b8e3302-utilities" (OuterVolumeSpecName: "utilities") pod "1451fb38-b06c-4b5c-bb43-8b571b8e3302" (UID: "1451fb38-b06c-4b5c-bb43-8b571b8e3302"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.404361 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1451fb38-b06c-4b5c-bb43-8b571b8e3302-kube-api-access-6zvjc" (OuterVolumeSpecName: "kube-api-access-6zvjc") pod "1451fb38-b06c-4b5c-bb43-8b571b8e3302" (UID: "1451fb38-b06c-4b5c-bb43-8b571b8e3302"). InnerVolumeSpecName "kube-api-access-6zvjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.418416 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1451fb38-b06c-4b5c-bb43-8b571b8e3302-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1451fb38-b06c-4b5c-bb43-8b571b8e3302" (UID: "1451fb38-b06c-4b5c-bb43-8b571b8e3302"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.500973 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zvjc\" (UniqueName: \"kubernetes.io/projected/1451fb38-b06c-4b5c-bb43-8b571b8e3302-kube-api-access-6zvjc\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.501026 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1451fb38-b06c-4b5c-bb43-8b571b8e3302-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:56 crc kubenswrapper[4629]: I1211 08:40:56.501041 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1451fb38-b06c-4b5c-bb43-8b571b8e3302-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:57 crc kubenswrapper[4629]: I1211 08:40:57.056421 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwgwp" Dec 11 08:40:57 crc kubenswrapper[4629]: I1211 08:40:57.056417 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwgwp" event={"ID":"1451fb38-b06c-4b5c-bb43-8b571b8e3302","Type":"ContainerDied","Data":"806c7656870463a3776ec16ac0ec6b46c9e1d96a5293bfa9f314e634a8c968b7"} Dec 11 08:40:57 crc kubenswrapper[4629]: I1211 08:40:57.056724 4629 scope.go:117] "RemoveContainer" containerID="33976ac375076b4955816ac01d2d04dac21238ee50ee5fed8940b8aeee10e8df" Dec 11 08:40:57 crc kubenswrapper[4629]: I1211 08:40:57.082376 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwgwp"] Dec 11 08:40:57 crc kubenswrapper[4629]: I1211 08:40:57.084158 4629 scope.go:117] "RemoveContainer" containerID="9f62ec2ab1a5d1dcdf6a3d4c53b894f9ef9306009cc6215146d8351366b9b42a" Dec 11 08:40:57 crc kubenswrapper[4629]: I1211 08:40:57.090680 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwgwp"] Dec 11 08:40:57 crc kubenswrapper[4629]: I1211 08:40:57.103029 4629 scope.go:117] "RemoveContainer" containerID="98a218ce29b1195393ad792517bab9a04c942ef264cd61b8775bd22ea187978a" Dec 11 08:40:58 crc kubenswrapper[4629]: I1211 08:40:58.206438 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" path="/var/lib/kubelet/pods/1451fb38-b06c-4b5c-bb43-8b571b8e3302/volumes" Dec 11 08:40:59 crc kubenswrapper[4629]: I1211 08:40:59.669000 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4jtwf"] Dec 11 08:41:02 crc kubenswrapper[4629]: I1211 08:41:02.149327 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:41:02 crc kubenswrapper[4629]: I1211 08:41:02.186681 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:41:04 crc kubenswrapper[4629]: I1211 08:41:04.337023 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6kzc"] Dec 11 08:41:04 crc kubenswrapper[4629]: I1211 08:41:04.337309 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c6kzc" podUID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" containerName="registry-server" containerID="cri-o://2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c" gracePeriod=2 Dec 11 08:41:04 crc kubenswrapper[4629]: I1211 08:41:04.729377 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:41:04 crc kubenswrapper[4629]: I1211 08:41:04.825227 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-utilities\") pod \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\" (UID: \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\") " Dec 11 08:41:04 crc kubenswrapper[4629]: I1211 08:41:04.825340 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-catalog-content\") pod \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\" (UID: \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\") " Dec 11 08:41:04 crc kubenswrapper[4629]: I1211 08:41:04.825388 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4qzq\" (UniqueName: \"kubernetes.io/projected/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-kube-api-access-v4qzq\") pod \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\" (UID: \"69c02574-1a57-4d6c-aa9c-8a85c6acd80f\") " Dec 11 08:41:04 crc kubenswrapper[4629]: I1211 08:41:04.827622 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-utilities" (OuterVolumeSpecName: "utilities") pod "69c02574-1a57-4d6c-aa9c-8a85c6acd80f" (UID: "69c02574-1a57-4d6c-aa9c-8a85c6acd80f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:41:04 crc kubenswrapper[4629]: I1211 08:41:04.846906 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-kube-api-access-v4qzq" (OuterVolumeSpecName: "kube-api-access-v4qzq") pod "69c02574-1a57-4d6c-aa9c-8a85c6acd80f" (UID: "69c02574-1a57-4d6c-aa9c-8a85c6acd80f"). InnerVolumeSpecName "kube-api-access-v4qzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:41:04 crc kubenswrapper[4629]: I1211 08:41:04.928232 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4qzq\" (UniqueName: \"kubernetes.io/projected/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-kube-api-access-v4qzq\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:04 crc kubenswrapper[4629]: I1211 08:41:04.928279 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:04 crc kubenswrapper[4629]: I1211 08:41:04.948172 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69c02574-1a57-4d6c-aa9c-8a85c6acd80f" (UID: "69c02574-1a57-4d6c-aa9c-8a85c6acd80f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.029342 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69c02574-1a57-4d6c-aa9c-8a85c6acd80f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.103448 4629 generic.go:334] "Generic (PLEG): container finished" podID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" containerID="2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c" exitCode=0 Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.103493 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kzc" event={"ID":"69c02574-1a57-4d6c-aa9c-8a85c6acd80f","Type":"ContainerDied","Data":"2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c"} Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.103545 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kzc" event={"ID":"69c02574-1a57-4d6c-aa9c-8a85c6acd80f","Type":"ContainerDied","Data":"1f890ae78abd2f0d10dcc514e3acef8d79a872c2cc7b2fd00ae6ea021d0bf081"} Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.103563 4629 scope.go:117] "RemoveContainer" containerID="2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.103568 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6kzc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.124900 4629 scope.go:117] "RemoveContainer" containerID="f4c1b5f7a2bc031a2d485d62b86b14fc29a06078466c8704b52d1776194588bd" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.134341 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6kzc"] Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.137585 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c6kzc"] Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.151780 4629 scope.go:117] "RemoveContainer" containerID="d1958a466fe1da0a31395861397896ea391f7ad02b8de9a319a6c61a6d0f87f3" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.176926 4629 scope.go:117] "RemoveContainer" containerID="2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.177560 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c\": container with ID starting with 2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c not found: ID does not exist" containerID="2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.177627 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c"} err="failed to get container status \"2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c\": rpc error: code = NotFound desc = could not find container \"2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c\": container with ID starting with 2c5fbb0d8252eae260e18851a714945fd23c33e1b2af5606b68c964c2343201c not found: ID does not exist" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.177656 4629 scope.go:117] "RemoveContainer" containerID="f4c1b5f7a2bc031a2d485d62b86b14fc29a06078466c8704b52d1776194588bd" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.178623 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c1b5f7a2bc031a2d485d62b86b14fc29a06078466c8704b52d1776194588bd\": container with ID starting with f4c1b5f7a2bc031a2d485d62b86b14fc29a06078466c8704b52d1776194588bd not found: ID does not exist" containerID="f4c1b5f7a2bc031a2d485d62b86b14fc29a06078466c8704b52d1776194588bd" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.178689 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c1b5f7a2bc031a2d485d62b86b14fc29a06078466c8704b52d1776194588bd"} err="failed to get container status \"f4c1b5f7a2bc031a2d485d62b86b14fc29a06078466c8704b52d1776194588bd\": rpc error: code = NotFound desc = could not find container \"f4c1b5f7a2bc031a2d485d62b86b14fc29a06078466c8704b52d1776194588bd\": container with ID starting with f4c1b5f7a2bc031a2d485d62b86b14fc29a06078466c8704b52d1776194588bd not found: ID does not exist" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.178717 4629 scope.go:117] "RemoveContainer" containerID="d1958a466fe1da0a31395861397896ea391f7ad02b8de9a319a6c61a6d0f87f3" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.179236 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1958a466fe1da0a31395861397896ea391f7ad02b8de9a319a6c61a6d0f87f3\": container with ID starting with d1958a466fe1da0a31395861397896ea391f7ad02b8de9a319a6c61a6d0f87f3 not found: ID does not exist" containerID="d1958a466fe1da0a31395861397896ea391f7ad02b8de9a319a6c61a6d0f87f3" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.179292 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1958a466fe1da0a31395861397896ea391f7ad02b8de9a319a6c61a6d0f87f3"} err="failed to get container status \"d1958a466fe1da0a31395861397896ea391f7ad02b8de9a319a6c61a6d0f87f3\": rpc error: code = NotFound desc = could not find container \"d1958a466fe1da0a31395861397896ea391f7ad02b8de9a319a6c61a6d0f87f3\": container with ID starting with d1958a466fe1da0a31395861397896ea391f7ad02b8de9a319a6c61a6d0f87f3 not found: ID does not exist" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.358762 4629 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.358821 4629 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359104 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" containerName="extract-utilities" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359117 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" containerName="extract-utilities" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359179 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" containerName="extract-content" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359187 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" containerName="extract-content" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359197 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" containerName="extract-content" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359203 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" containerName="extract-content" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359213 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359220 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359226 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b0de2ef-6a6c-460c-bf25-325cee52f44f" containerName="pruner" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359231 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b0de2ef-6a6c-460c-bf25-325cee52f44f" containerName="pruner" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359241 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359246 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359253 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" containerName="extract-utilities" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359259 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" containerName="extract-utilities" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359267 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359274 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359282 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359287 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359295 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" containerName="extract-utilities" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359323 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" containerName="extract-utilities" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359332 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" containerName="registry-server" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359338 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" containerName="registry-server" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359346 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359352 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359359 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" containerName="registry-server" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359364 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" containerName="registry-server" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359371 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359377 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359383 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" containerName="registry-server" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359389 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" containerName="registry-server" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359399 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" containerName="extract-content" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359404 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" containerName="extract-content" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359413 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" containerName="extract-content" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359420 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" containerName="extract-content" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359429 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" containerName="extract-utilities" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359435 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" containerName="extract-utilities" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359443 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359448 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 11 08:41:05 crc kubenswrapper[4629]: E1211 08:41:05.359457 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" containerName="registry-server" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359462 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" containerName="registry-server" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359544 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b0de2ef-6a6c-460c-bf25-325cee52f44f" containerName="pruner" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359551 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" containerName="registry-server" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359560 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359569 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359577 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359584 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359592 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359598 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="7adee9e0-edb2-4ea3-afa6-d5759440a79a" containerName="registry-server" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359605 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="1451fb38-b06c-4b5c-bb43-8b571b8e3302" containerName="registry-server" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359612 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.359618 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cc54f96-ce8c-4efe-9c1f-ac612c9ab366" containerName="registry-server" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.362856 4629 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.363317 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.366990 4629 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.395085 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.534996 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.535039 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.535084 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.535119 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.535144 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.535170 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.535185 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.535201 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638140 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638219 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638248 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638278 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638294 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638289 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638340 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638370 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638380 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638310 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638351 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638314 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638458 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638481 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638568 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.638581 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: I1211 08:41:05.692630 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:05 crc kubenswrapper[4629]: W1211 08:41:05.713591 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-b3a750c1e0454cd160021fd7e9cdee96b0aa7207f05efd22860ea2e77e185ca7 WatchSource:0}: Error finding container b3a750c1e0454cd160021fd7e9cdee96b0aa7207f05efd22860ea2e77e185ca7: Status 404 returned error can't find the container with id b3a750c1e0454cd160021fd7e9cdee96b0aa7207f05efd22860ea2e77e185ca7 Dec 11 08:41:06 crc kubenswrapper[4629]: I1211 08:41:06.109599 4629 generic.go:334] "Generic (PLEG): container finished" podID="695de9c8-1898-48e1-8967-982a5d2cb3ab" containerID="fe20e691db6497318b74c87a3fd4b77dcd02b5e8f999b7624f94e861f7ec1e74" exitCode=0 Dec 11 08:41:06 crc kubenswrapper[4629]: I1211 08:41:06.109688 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"695de9c8-1898-48e1-8967-982a5d2cb3ab","Type":"ContainerDied","Data":"fe20e691db6497318b74c87a3fd4b77dcd02b5e8f999b7624f94e861f7ec1e74"} Dec 11 08:41:06 crc kubenswrapper[4629]: I1211 08:41:06.112592 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"14cd17fa30381fcce30a71be2280881a94bb1e28df0b235f4ddc4af5c89acd12"} Dec 11 08:41:06 crc kubenswrapper[4629]: I1211 08:41:06.112655 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b3a750c1e0454cd160021fd7e9cdee96b0aa7207f05efd22860ea2e77e185ca7"} Dec 11 08:41:06 crc kubenswrapper[4629]: I1211 08:41:06.115250 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468" gracePeriod=15 Dec 11 08:41:06 crc kubenswrapper[4629]: I1211 08:41:06.115428 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da" gracePeriod=15 Dec 11 08:41:06 crc kubenswrapper[4629]: I1211 08:41:06.115479 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4" gracePeriod=15 Dec 11 08:41:06 crc kubenswrapper[4629]: I1211 08:41:06.115525 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d" gracePeriod=15 Dec 11 08:41:06 crc kubenswrapper[4629]: I1211 08:41:06.115567 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd" gracePeriod=15 Dec 11 08:41:06 crc kubenswrapper[4629]: I1211 08:41:06.211912 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69c02574-1a57-4d6c-aa9c-8a85c6acd80f" path="/var/lib/kubelet/pods/69c02574-1a57-4d6c-aa9c-8a85c6acd80f/volumes" Dec 11 08:41:06 crc kubenswrapper[4629]: E1211 08:41:06.977685 4629 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:06 crc kubenswrapper[4629]: E1211 08:41:06.978121 4629 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:06 crc kubenswrapper[4629]: E1211 08:41:06.978321 4629 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:06 crc kubenswrapper[4629]: E1211 08:41:06.978467 4629 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:06 crc kubenswrapper[4629]: E1211 08:41:06.978609 4629 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:06 crc kubenswrapper[4629]: I1211 08:41:06.978637 4629 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 11 08:41:06 crc kubenswrapper[4629]: E1211 08:41:06.978766 4629 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" interval="200ms" Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.129228 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.130696 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.131421 4629 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da" exitCode=0 Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.131453 4629 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4" exitCode=0 Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.131466 4629 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d" exitCode=0 Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.131475 4629 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd" exitCode=2 Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.131656 4629 scope.go:117] "RemoveContainer" containerID="f3c17a84324bfbed67d993efd5633f644abfcd6652fc7cd1ea161bd970fbf6fa" Dec 11 08:41:07 crc kubenswrapper[4629]: E1211 08:41:07.180467 4629 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" interval="400ms" Dec 11 08:41:07 crc kubenswrapper[4629]: E1211 08:41:07.270595 4629 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.129.56.111:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" volumeName="registry-storage" Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.382767 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.570185 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/695de9c8-1898-48e1-8967-982a5d2cb3ab-var-lock\") pod \"695de9c8-1898-48e1-8967-982a5d2cb3ab\" (UID: \"695de9c8-1898-48e1-8967-982a5d2cb3ab\") " Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.570282 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/695de9c8-1898-48e1-8967-982a5d2cb3ab-kubelet-dir\") pod \"695de9c8-1898-48e1-8967-982a5d2cb3ab\" (UID: \"695de9c8-1898-48e1-8967-982a5d2cb3ab\") " Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.570320 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/695de9c8-1898-48e1-8967-982a5d2cb3ab-kube-api-access\") pod \"695de9c8-1898-48e1-8967-982a5d2cb3ab\" (UID: \"695de9c8-1898-48e1-8967-982a5d2cb3ab\") " Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.570328 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/695de9c8-1898-48e1-8967-982a5d2cb3ab-var-lock" (OuterVolumeSpecName: "var-lock") pod "695de9c8-1898-48e1-8967-982a5d2cb3ab" (UID: "695de9c8-1898-48e1-8967-982a5d2cb3ab"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.570398 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/695de9c8-1898-48e1-8967-982a5d2cb3ab-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "695de9c8-1898-48e1-8967-982a5d2cb3ab" (UID: "695de9c8-1898-48e1-8967-982a5d2cb3ab"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.570560 4629 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/695de9c8-1898-48e1-8967-982a5d2cb3ab-var-lock\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.570573 4629 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/695de9c8-1898-48e1-8967-982a5d2cb3ab-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:07 crc kubenswrapper[4629]: E1211 08:41:07.581313 4629 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" interval="800ms" Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.586253 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/695de9c8-1898-48e1-8967-982a5d2cb3ab-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "695de9c8-1898-48e1-8967-982a5d2cb3ab" (UID: "695de9c8-1898-48e1-8967-982a5d2cb3ab"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:41:07 crc kubenswrapper[4629]: I1211 08:41:07.671829 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/695de9c8-1898-48e1-8967-982a5d2cb3ab-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.138189 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.141160 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"695de9c8-1898-48e1-8967-982a5d2cb3ab","Type":"ContainerDied","Data":"eef351cd94a652798e5e93e09b019efc3cdb21d0b2a857a797d57ce7b322ef98"} Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.141221 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eef351cd94a652798e5e93e09b019efc3cdb21d0b2a857a797d57ce7b322ef98" Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.145039 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 08:41:08 crc kubenswrapper[4629]: E1211 08:41:08.382568 4629 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" interval="1.6s" Dec 11 08:41:08 crc kubenswrapper[4629]: E1211 08:41:08.580021 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:41:08Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:41:08Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:41:08Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:41:08Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:08 crc kubenswrapper[4629]: E1211 08:41:08.580228 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:08 crc kubenswrapper[4629]: E1211 08:41:08.580374 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:08 crc kubenswrapper[4629]: E1211 08:41:08.580513 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:08 crc kubenswrapper[4629]: E1211 08:41:08.580664 4629 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:08 crc kubenswrapper[4629]: E1211 08:41:08.580679 4629 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.626943 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.628047 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.784068 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.784141 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.784178 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.784244 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.784244 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.784387 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.784602 4629 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.784623 4629 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:08 crc kubenswrapper[4629]: I1211 08:41:08.784638 4629 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.155087 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.156419 4629 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468" exitCode=0 Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.156484 4629 scope.go:117] "RemoveContainer" containerID="03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.156648 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.186620 4629 scope.go:117] "RemoveContainer" containerID="defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.201217 4629 scope.go:117] "RemoveContainer" containerID="ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.214678 4629 scope.go:117] "RemoveContainer" containerID="d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.228156 4629 scope.go:117] "RemoveContainer" containerID="b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.247609 4629 scope.go:117] "RemoveContainer" containerID="e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.279147 4629 scope.go:117] "RemoveContainer" containerID="03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da" Dec 11 08:41:09 crc kubenswrapper[4629]: E1211 08:41:09.280387 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\": container with ID starting with 03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da not found: ID does not exist" containerID="03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.280432 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da"} err="failed to get container status \"03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\": rpc error: code = NotFound desc = could not find container \"03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da\": container with ID starting with 03497880f218501d312c058b23b28d8d2520497011fd9c5a549d7fde6ffd11da not found: ID does not exist" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.280460 4629 scope.go:117] "RemoveContainer" containerID="defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4" Dec 11 08:41:09 crc kubenswrapper[4629]: E1211 08:41:09.280812 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\": container with ID starting with defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4 not found: ID does not exist" containerID="defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.280829 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4"} err="failed to get container status \"defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\": rpc error: code = NotFound desc = could not find container \"defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4\": container with ID starting with defa1f5f0204344f3404862523cdcce3ecab4e172a6e1224c5f67f5c797cf0c4 not found: ID does not exist" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.280842 4629 scope.go:117] "RemoveContainer" containerID="ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d" Dec 11 08:41:09 crc kubenswrapper[4629]: E1211 08:41:09.281249 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\": container with ID starting with ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d not found: ID does not exist" containerID="ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.281290 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d"} err="failed to get container status \"ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\": rpc error: code = NotFound desc = could not find container \"ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d\": container with ID starting with ec07ce9171d62cf6fba3726227971096fa7907d10b37f7ee3b5ec6a025fa004d not found: ID does not exist" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.281307 4629 scope.go:117] "RemoveContainer" containerID="d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd" Dec 11 08:41:09 crc kubenswrapper[4629]: E1211 08:41:09.281694 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\": container with ID starting with d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd not found: ID does not exist" containerID="d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.281728 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd"} err="failed to get container status \"d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\": rpc error: code = NotFound desc = could not find container \"d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd\": container with ID starting with d4fd080f908903afe688637f7d633c1ee80579b1224685529291a554870e75dd not found: ID does not exist" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.281756 4629 scope.go:117] "RemoveContainer" containerID="b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468" Dec 11 08:41:09 crc kubenswrapper[4629]: E1211 08:41:09.282540 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\": container with ID starting with b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468 not found: ID does not exist" containerID="b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.282578 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468"} err="failed to get container status \"b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\": rpc error: code = NotFound desc = could not find container \"b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468\": container with ID starting with b42b2008787a76edb4abb6ac65970315db6b6718c98cc4c0147214e02a5e3468 not found: ID does not exist" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.282604 4629 scope.go:117] "RemoveContainer" containerID="e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838" Dec 11 08:41:09 crc kubenswrapper[4629]: E1211 08:41:09.283114 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\": container with ID starting with e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838 not found: ID does not exist" containerID="e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838" Dec 11 08:41:09 crc kubenswrapper[4629]: I1211 08:41:09.283143 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838"} err="failed to get container status \"e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\": rpc error: code = NotFound desc = could not find container \"e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838\": container with ID starting with e50fe81e71c7d64e55c742e25e32cdce34181abb5eced3fecead53c69b034838 not found: ID does not exist" Dec 11 08:41:09 crc kubenswrapper[4629]: E1211 08:41:09.983808 4629 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" interval="3.2s" Dec 11 08:41:10 crc kubenswrapper[4629]: I1211 08:41:10.209112 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 11 08:41:11 crc kubenswrapper[4629]: I1211 08:41:11.181262 4629 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:11 crc kubenswrapper[4629]: I1211 08:41:11.181811 4629 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:11 crc kubenswrapper[4629]: I1211 08:41:11.182274 4629 status_manager.go:851] "Failed to get status for pod" podUID="695de9c8-1898-48e1-8967-982a5d2cb3ab" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:11 crc kubenswrapper[4629]: I1211 08:41:11.182659 4629 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:12 crc kubenswrapper[4629]: I1211 08:41:12.201581 4629 status_manager.go:851] "Failed to get status for pod" podUID="695de9c8-1898-48e1-8967-982a5d2cb3ab" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:12 crc kubenswrapper[4629]: I1211 08:41:12.204273 4629 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:13 crc kubenswrapper[4629]: E1211 08:41:13.185537 4629 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.111:6443: connect: connection refused" interval="6.4s" Dec 11 08:41:16 crc kubenswrapper[4629]: I1211 08:41:16.198413 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:16 crc kubenswrapper[4629]: I1211 08:41:16.200550 4629 status_manager.go:851] "Failed to get status for pod" podUID="695de9c8-1898-48e1-8967-982a5d2cb3ab" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:16 crc kubenswrapper[4629]: I1211 08:41:16.201316 4629 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:16 crc kubenswrapper[4629]: I1211 08:41:16.217766 4629 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="37701d38-a859-4d67-9257-a054c616bc26" Dec 11 08:41:16 crc kubenswrapper[4629]: I1211 08:41:16.217835 4629 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="37701d38-a859-4d67-9257-a054c616bc26" Dec 11 08:41:16 crc kubenswrapper[4629]: E1211 08:41:16.218428 4629 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.111:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:16 crc kubenswrapper[4629]: I1211 08:41:16.218905 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:16 crc kubenswrapper[4629]: E1211 08:41:16.244729 4629 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.111:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-crc.18801c9413e54ab0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:71bb4a3aecc4ba5b26c4b7318770ce13,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 08:41:16.243765936 +0000 UTC m=+264.309181544,LastTimestamp:2025-12-11 08:41:16.243765936 +0000 UTC m=+264.309181544,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 08:41:17 crc kubenswrapper[4629]: I1211 08:41:17.196698 4629 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="cca675ad22e5eb840c3d7bea557c0d7e34a84d0ac5a02e99cc13c97e1305e29b" exitCode=0 Dec 11 08:41:17 crc kubenswrapper[4629]: I1211 08:41:17.196755 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"cca675ad22e5eb840c3d7bea557c0d7e34a84d0ac5a02e99cc13c97e1305e29b"} Dec 11 08:41:17 crc kubenswrapper[4629]: I1211 08:41:17.196788 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3051fdae95ff0da2e408da5863d32bfb2fbe08efde825953a3eeb46efc3ba60f"} Dec 11 08:41:17 crc kubenswrapper[4629]: I1211 08:41:17.197155 4629 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="37701d38-a859-4d67-9257-a054c616bc26" Dec 11 08:41:17 crc kubenswrapper[4629]: I1211 08:41:17.197180 4629 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="37701d38-a859-4d67-9257-a054c616bc26" Dec 11 08:41:17 crc kubenswrapper[4629]: E1211 08:41:17.197548 4629 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.111:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:17 crc kubenswrapper[4629]: I1211 08:41:17.198058 4629 status_manager.go:851] "Failed to get status for pod" podUID="695de9c8-1898-48e1-8967-982a5d2cb3ab" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:17 crc kubenswrapper[4629]: I1211 08:41:17.198402 4629 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.111:6443: connect: connection refused" Dec 11 08:41:18 crc kubenswrapper[4629]: I1211 08:41:18.217099 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"526f01ac13fd32840ddbf5f9482ba817878f78d642a226c55506120483e96f10"} Dec 11 08:41:18 crc kubenswrapper[4629]: I1211 08:41:18.217478 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"73bbc198e68d5627407c412a34b341472c0b14ecfd5d4b80df33bf6ab8b100e1"} Dec 11 08:41:18 crc kubenswrapper[4629]: I1211 08:41:18.217493 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d9c70f21840394fd216dce37c576d5160537c0a5e251c9920026088c30d99d1e"} Dec 11 08:41:18 crc kubenswrapper[4629]: I1211 08:41:18.217505 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4deee5cefe58b1b77420197dfb9c21dd2dd8238d7b474ae6490dfb55c833558c"} Dec 11 08:41:19 crc kubenswrapper[4629]: I1211 08:41:19.218697 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4633b25505373b810eea1977719e2283fd3b96f63d8f8d6a7732b0a9bbb7d202"} Dec 11 08:41:19 crc kubenswrapper[4629]: I1211 08:41:19.219093 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:19 crc kubenswrapper[4629]: I1211 08:41:19.220236 4629 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="37701d38-a859-4d67-9257-a054c616bc26" Dec 11 08:41:19 crc kubenswrapper[4629]: I1211 08:41:19.220351 4629 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="37701d38-a859-4d67-9257-a054c616bc26" Dec 11 08:41:20 crc kubenswrapper[4629]: I1211 08:41:20.225573 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 11 08:41:20 crc kubenswrapper[4629]: I1211 08:41:20.225617 4629 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4" exitCode=1 Dec 11 08:41:20 crc kubenswrapper[4629]: I1211 08:41:20.225644 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4"} Dec 11 08:41:20 crc kubenswrapper[4629]: I1211 08:41:20.226056 4629 scope.go:117] "RemoveContainer" containerID="089e91d9b988baebf0a6febbd2dcedd75ceaf0e35c845dd3c99cc459f92c6eb4" Dec 11 08:41:21 crc kubenswrapper[4629]: I1211 08:41:21.219791 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:21 crc kubenswrapper[4629]: I1211 08:41:21.220471 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:21 crc kubenswrapper[4629]: I1211 08:41:21.228181 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:21 crc kubenswrapper[4629]: I1211 08:41:21.233507 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 11 08:41:21 crc kubenswrapper[4629]: I1211 08:41:21.234004 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bdd349b096d0e57f727d7f1dd76ab1819a10a2fad5a7f46eff192c5c8e7b3c56"} Dec 11 08:41:24 crc kubenswrapper[4629]: I1211 08:41:24.255877 4629 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:24 crc kubenswrapper[4629]: I1211 08:41:24.319138 4629 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="fe6fbff0-4377-4225-b3d6-76a4a35dd0f5" Dec 11 08:41:24 crc kubenswrapper[4629]: I1211 08:41:24.693023 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" podUID="449eb3bb-7054-452d-85d6-558aa5efa682" containerName="oauth-openshift" containerID="cri-o://94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059" gracePeriod=15 Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.078990 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.167710 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-ocp-branding-template\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.167764 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-service-ca\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.167799 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/449eb3bb-7054-452d-85d6-558aa5efa682-audit-dir\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.167825 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-error\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.167890 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85h6k\" (UniqueName: \"kubernetes.io/projected/449eb3bb-7054-452d-85d6-558aa5efa682-kube-api-access-85h6k\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.167917 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-idp-0-file-data\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.167910 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449eb3bb-7054-452d-85d6-558aa5efa682-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.167943 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-session\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.167976 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-provider-selection\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.168006 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-trusted-ca-bundle\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.168021 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-audit-policies\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.168096 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-login\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.168129 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-serving-cert\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.168145 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-cliconfig\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.168176 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-router-certs\") pod \"449eb3bb-7054-452d-85d6-558aa5efa682\" (UID: \"449eb3bb-7054-452d-85d6-558aa5efa682\") " Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.168642 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.169000 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.169020 4629 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/449eb3bb-7054-452d-85d6-558aa5efa682-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.174035 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.174607 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.174862 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.174986 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.175287 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.175556 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.175579 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.175990 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.177009 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.183058 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/449eb3bb-7054-452d-85d6-558aa5efa682-kube-api-access-85h6k" (OuterVolumeSpecName: "kube-api-access-85h6k") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "kube-api-access-85h6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.183932 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.184338 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "449eb3bb-7054-452d-85d6-558aa5efa682" (UID: "449eb3bb-7054-452d-85d6-558aa5efa682"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.255469 4629 generic.go:334] "Generic (PLEG): container finished" podID="449eb3bb-7054-452d-85d6-558aa5efa682" containerID="94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059" exitCode=0 Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.255555 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.255654 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" event={"ID":"449eb3bb-7054-452d-85d6-558aa5efa682","Type":"ContainerDied","Data":"94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059"} Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.255681 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4jtwf" event={"ID":"449eb3bb-7054-452d-85d6-558aa5efa682","Type":"ContainerDied","Data":"4da6ae0bfa62287ccc5e9a0da29741bb282b3cb062a462e63ed1e58610ef2f6d"} Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.255697 4629 scope.go:117] "RemoveContainer" containerID="94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.256284 4629 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="37701d38-a859-4d67-9257-a054c616bc26" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.256297 4629 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="37701d38-a859-4d67-9257-a054c616bc26" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.263196 4629 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="fe6fbff0-4377-4225-b3d6-76a4a35dd0f5" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.269889 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.270107 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.270201 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.270274 4629 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.270337 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.270396 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.270455 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.270517 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.270606 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.270664 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.270721 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85h6k\" (UniqueName: \"kubernetes.io/projected/449eb3bb-7054-452d-85d6-558aa5efa682-kube-api-access-85h6k\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.270786 4629 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/449eb3bb-7054-452d-85d6-558aa5efa682-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.283582 4629 scope.go:117] "RemoveContainer" containerID="94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059" Dec 11 08:41:25 crc kubenswrapper[4629]: E1211 08:41:25.284120 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059\": container with ID starting with 94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059 not found: ID does not exist" containerID="94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059" Dec 11 08:41:25 crc kubenswrapper[4629]: I1211 08:41:25.284150 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059"} err="failed to get container status \"94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059\": rpc error: code = NotFound desc = could not find container \"94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059\": container with ID starting with 94aae7581bb8d1f89e10b6d9059a9d0973ec3706bd6fba765c0b436fa6bc4059 not found: ID does not exist" Dec 11 08:41:26 crc kubenswrapper[4629]: I1211 08:41:26.258918 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:41:26 crc kubenswrapper[4629]: I1211 08:41:26.883632 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:41:26 crc kubenswrapper[4629]: I1211 08:41:26.887575 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:41:31 crc kubenswrapper[4629]: I1211 08:41:31.109053 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 11 08:41:31 crc kubenswrapper[4629]: I1211 08:41:31.170336 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 11 08:41:31 crc kubenswrapper[4629]: I1211 08:41:31.171529 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 11 08:41:31 crc kubenswrapper[4629]: I1211 08:41:31.286063 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 11 08:41:31 crc kubenswrapper[4629]: I1211 08:41:31.293058 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 11 08:41:31 crc kubenswrapper[4629]: I1211 08:41:31.329446 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 11 08:41:31 crc kubenswrapper[4629]: I1211 08:41:31.393543 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 11 08:41:31 crc kubenswrapper[4629]: I1211 08:41:31.540643 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 11 08:41:31 crc kubenswrapper[4629]: I1211 08:41:31.772562 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 11 08:41:31 crc kubenswrapper[4629]: I1211 08:41:31.924028 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 11 08:41:31 crc kubenswrapper[4629]: I1211 08:41:31.987955 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 11 08:41:32 crc kubenswrapper[4629]: I1211 08:41:32.207433 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 11 08:41:32 crc kubenswrapper[4629]: I1211 08:41:32.305287 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 11 08:41:32 crc kubenswrapper[4629]: I1211 08:41:32.309224 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 11 08:41:32 crc kubenswrapper[4629]: I1211 08:41:32.363867 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 11 08:41:32 crc kubenswrapper[4629]: I1211 08:41:32.452260 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 11 08:41:32 crc kubenswrapper[4629]: I1211 08:41:32.459464 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 11 08:41:32 crc kubenswrapper[4629]: I1211 08:41:32.488541 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 11 08:41:32 crc kubenswrapper[4629]: I1211 08:41:32.496896 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 11 08:41:32 crc kubenswrapper[4629]: I1211 08:41:32.676141 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 11 08:41:32 crc kubenswrapper[4629]: I1211 08:41:32.976940 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 11 08:41:33 crc kubenswrapper[4629]: I1211 08:41:33.003296 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 11 08:41:33 crc kubenswrapper[4629]: I1211 08:41:33.112898 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 11 08:41:33 crc kubenswrapper[4629]: I1211 08:41:33.123187 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 11 08:41:33 crc kubenswrapper[4629]: I1211 08:41:33.129868 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 11 08:41:33 crc kubenswrapper[4629]: I1211 08:41:33.167578 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 11 08:41:33 crc kubenswrapper[4629]: I1211 08:41:33.240581 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 11 08:41:33 crc kubenswrapper[4629]: I1211 08:41:33.494742 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 11 08:41:33 crc kubenswrapper[4629]: I1211 08:41:33.850066 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 11 08:41:33 crc kubenswrapper[4629]: I1211 08:41:33.999781 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.031586 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.201591 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.281624 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.299432 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.322589 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.361387 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.365228 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.382624 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.609056 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.693081 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.733515 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.802182 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 11 08:41:34 crc kubenswrapper[4629]: I1211 08:41:34.821875 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 11 08:41:35 crc kubenswrapper[4629]: I1211 08:41:35.229312 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 11 08:41:35 crc kubenswrapper[4629]: I1211 08:41:35.311534 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 11 08:41:35 crc kubenswrapper[4629]: I1211 08:41:35.323042 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 11 08:41:35 crc kubenswrapper[4629]: I1211 08:41:35.466118 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 11 08:41:35 crc kubenswrapper[4629]: I1211 08:41:35.561639 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 11 08:41:35 crc kubenswrapper[4629]: I1211 08:41:35.780705 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 11 08:41:35 crc kubenswrapper[4629]: I1211 08:41:35.860400 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 11 08:41:35 crc kubenswrapper[4629]: I1211 08:41:35.956593 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 11 08:41:35 crc kubenswrapper[4629]: I1211 08:41:35.977120 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 11 08:41:36 crc kubenswrapper[4629]: I1211 08:41:36.017712 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 11 08:41:36 crc kubenswrapper[4629]: I1211 08:41:36.030706 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 11 08:41:36 crc kubenswrapper[4629]: I1211 08:41:36.166784 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 11 08:41:36 crc kubenswrapper[4629]: I1211 08:41:36.207532 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 11 08:41:36 crc kubenswrapper[4629]: I1211 08:41:36.262743 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:41:36 crc kubenswrapper[4629]: I1211 08:41:36.442891 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 11 08:41:36 crc kubenswrapper[4629]: I1211 08:41:36.474329 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 11 08:41:36 crc kubenswrapper[4629]: I1211 08:41:36.637707 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 11 08:41:36 crc kubenswrapper[4629]: I1211 08:41:36.729957 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 11 08:41:36 crc kubenswrapper[4629]: I1211 08:41:36.803428 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 11 08:41:36 crc kubenswrapper[4629]: I1211 08:41:36.906654 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 11 08:41:37 crc kubenswrapper[4629]: I1211 08:41:37.321260 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 11 08:41:37 crc kubenswrapper[4629]: I1211 08:41:37.505826 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 11 08:41:37 crc kubenswrapper[4629]: I1211 08:41:37.661935 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 11 08:41:38 crc kubenswrapper[4629]: I1211 08:41:38.097508 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 11 08:41:38 crc kubenswrapper[4629]: I1211 08:41:38.108026 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 11 08:41:38 crc kubenswrapper[4629]: I1211 08:41:38.189691 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 11 08:41:38 crc kubenswrapper[4629]: I1211 08:41:38.388567 4629 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 11 08:41:38 crc kubenswrapper[4629]: I1211 08:41:38.400599 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 11 08:41:38 crc kubenswrapper[4629]: I1211 08:41:38.431580 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 11 08:41:38 crc kubenswrapper[4629]: I1211 08:41:38.790564 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 11 08:41:39 crc kubenswrapper[4629]: I1211 08:41:39.173811 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 11 08:41:39 crc kubenswrapper[4629]: I1211 08:41:39.194967 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 11 08:41:39 crc kubenswrapper[4629]: I1211 08:41:39.222665 4629 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 11 08:41:39 crc kubenswrapper[4629]: I1211 08:41:39.280460 4629 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 11 08:41:39 crc kubenswrapper[4629]: I1211 08:41:39.312229 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 11 08:41:39 crc kubenswrapper[4629]: I1211 08:41:39.671378 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 11 08:41:39 crc kubenswrapper[4629]: I1211 08:41:39.949010 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 11 08:41:39 crc kubenswrapper[4629]: I1211 08:41:39.971958 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.133885 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.151953 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.228449 4629 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.233091 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=35.233030329 podStartE2EDuration="35.233030329s" podCreationTimestamp="2025-12-11 08:41:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:41:24.285808814 +0000 UTC m=+272.351224432" watchObservedRunningTime="2025-12-11 08:41:40.233030329 +0000 UTC m=+288.298445937" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.234874 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4jtwf","openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.234959 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.240651 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.240721 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.322946 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.322915821 podStartE2EDuration="16.322915821s" podCreationTimestamp="2025-12-11 08:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:41:40.260191873 +0000 UTC m=+288.325607491" watchObservedRunningTime="2025-12-11 08:41:40.322915821 +0000 UTC m=+288.388331429" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.388547 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.417066 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.422411 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.585166 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.654380 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.676985 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.757687 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.799057 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.800905 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.815204 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 11 08:41:40 crc kubenswrapper[4629]: I1211 08:41:40.987177 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 11 08:41:41 crc kubenswrapper[4629]: I1211 08:41:41.206917 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 11 08:41:41 crc kubenswrapper[4629]: I1211 08:41:41.571656 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 11 08:41:41 crc kubenswrapper[4629]: I1211 08:41:41.573806 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 11 08:41:41 crc kubenswrapper[4629]: I1211 08:41:41.625699 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 11 08:41:41 crc kubenswrapper[4629]: I1211 08:41:41.795769 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 11 08:41:41 crc kubenswrapper[4629]: I1211 08:41:41.813512 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 11 08:41:41 crc kubenswrapper[4629]: I1211 08:41:41.838268 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 11 08:41:41 crc kubenswrapper[4629]: I1211 08:41:41.943027 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 11 08:41:41 crc kubenswrapper[4629]: I1211 08:41:41.945753 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 11 08:41:41 crc kubenswrapper[4629]: I1211 08:41:41.979634 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 11 08:41:41 crc kubenswrapper[4629]: I1211 08:41:41.994406 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.206120 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="449eb3bb-7054-452d-85d6-558aa5efa682" path="/var/lib/kubelet/pods/449eb3bb-7054-452d-85d6-558aa5efa682/volumes" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.226997 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.302089 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.412668 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.513550 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.543831 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.616331 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.654779 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.679173 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.698500 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.716341 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.733666 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.843773 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.977772 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 11 08:41:42 crc kubenswrapper[4629]: I1211 08:41:42.998818 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 11 08:41:43 crc kubenswrapper[4629]: I1211 08:41:43.048613 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 11 08:41:43 crc kubenswrapper[4629]: I1211 08:41:43.071705 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 11 08:41:43 crc kubenswrapper[4629]: I1211 08:41:43.073711 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 11 08:41:43 crc kubenswrapper[4629]: I1211 08:41:43.271094 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 11 08:41:43 crc kubenswrapper[4629]: I1211 08:41:43.414840 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 11 08:41:43 crc kubenswrapper[4629]: I1211 08:41:43.457280 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 11 08:41:43 crc kubenswrapper[4629]: I1211 08:41:43.516293 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 11 08:41:43 crc kubenswrapper[4629]: I1211 08:41:43.719033 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 11 08:41:43 crc kubenswrapper[4629]: I1211 08:41:43.753398 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 11 08:41:43 crc kubenswrapper[4629]: I1211 08:41:43.787571 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 11 08:41:43 crc kubenswrapper[4629]: I1211 08:41:43.904942 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.081771 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.263480 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.278187 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.281165 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.377138 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.383580 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.402959 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.414119 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.431490 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.456022 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.497956 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.556394 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.664717 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.681545 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.897956 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.907045 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.989340 4629 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 11 08:41:44 crc kubenswrapper[4629]: I1211 08:41:44.994113 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.015677 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.080442 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.193888 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.216160 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.234750 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.376884 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.462247 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.622024 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.764583 4629 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.764845 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://14cd17fa30381fcce30a71be2280881a94bb1e28df0b235f4ddc4af5c89acd12" gracePeriod=5 Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.793968 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.953575 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 11 08:41:45 crc kubenswrapper[4629]: I1211 08:41:45.983314 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 11 08:41:46 crc kubenswrapper[4629]: I1211 08:41:46.051486 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 11 08:41:46 crc kubenswrapper[4629]: I1211 08:41:46.113152 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 11 08:41:46 crc kubenswrapper[4629]: I1211 08:41:46.119509 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 11 08:41:46 crc kubenswrapper[4629]: I1211 08:41:46.181900 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 11 08:41:46 crc kubenswrapper[4629]: I1211 08:41:46.361421 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 11 08:41:46 crc kubenswrapper[4629]: I1211 08:41:46.394196 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 11 08:41:46 crc kubenswrapper[4629]: I1211 08:41:46.649455 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 11 08:41:46 crc kubenswrapper[4629]: I1211 08:41:46.763454 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 11 08:41:46 crc kubenswrapper[4629]: I1211 08:41:46.778169 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 11 08:41:46 crc kubenswrapper[4629]: I1211 08:41:46.823982 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 11 08:41:46 crc kubenswrapper[4629]: I1211 08:41:46.876774 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 11 08:41:46 crc kubenswrapper[4629]: I1211 08:41:46.889656 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.015921 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.016378 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.022630 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.055778 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.153607 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.185224 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.319321 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.426480 4629 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.478869 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.653124 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.664543 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.694824 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.789076 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.823546 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 11 08:41:47 crc kubenswrapper[4629]: I1211 08:41:47.955043 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.039228 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.063600 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.250815 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.276347 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.281161 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.303932 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5db8794bf8-j99ff"] Dec 11 08:41:48 crc kubenswrapper[4629]: E1211 08:41:48.304158 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449eb3bb-7054-452d-85d6-558aa5efa682" containerName="oauth-openshift" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.304171 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="449eb3bb-7054-452d-85d6-558aa5efa682" containerName="oauth-openshift" Dec 11 08:41:48 crc kubenswrapper[4629]: E1211 08:41:48.304182 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="695de9c8-1898-48e1-8967-982a5d2cb3ab" containerName="installer" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.304189 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="695de9c8-1898-48e1-8967-982a5d2cb3ab" containerName="installer" Dec 11 08:41:48 crc kubenswrapper[4629]: E1211 08:41:48.304208 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.304216 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.304327 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="449eb3bb-7054-452d-85d6-558aa5efa682" containerName="oauth-openshift" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.304340 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="695de9c8-1898-48e1-8967-982a5d2cb3ab" containerName="installer" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.304352 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.304740 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.310981 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.311806 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.311896 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.311993 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.312256 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.312415 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.312557 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.312693 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.312823 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.313130 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.313641 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.313643 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.313963 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.325006 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.328239 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.332695 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.334811 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.336881 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5db8794bf8-j99ff"] Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.369493 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.466119 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-service-ca\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.466825 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-router-certs\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.467001 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/96cf26ec-8a64-472c-86a5-918b5170ce61-audit-dir\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.467098 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.467209 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.467363 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.467508 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.467635 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-session\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.467757 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fgzq\" (UniqueName: \"kubernetes.io/projected/96cf26ec-8a64-472c-86a5-918b5170ce61-kube-api-access-5fgzq\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.467884 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/96cf26ec-8a64-472c-86a5-918b5170ce61-audit-policies\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.467990 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-user-template-error\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.468111 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-user-template-login\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.468221 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.468314 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.511556 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.539318 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.543388 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.567321 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.569821 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-service-ca\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.569963 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-router-certs\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.570059 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/96cf26ec-8a64-472c-86a5-918b5170ce61-audit-dir\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.570174 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.570269 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.570356 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.570450 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.570527 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-session\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.570608 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fgzq\" (UniqueName: \"kubernetes.io/projected/96cf26ec-8a64-472c-86a5-918b5170ce61-kube-api-access-5fgzq\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.570702 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/96cf26ec-8a64-472c-86a5-918b5170ce61-audit-policies\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.570782 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-user-template-error\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.570958 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-user-template-login\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.571048 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.571126 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.572890 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/96cf26ec-8a64-472c-86a5-918b5170ce61-audit-policies\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.576544 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.576596 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-user-template-login\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.577350 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-service-ca\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.579612 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-user-template-error\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.579953 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.580009 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/96cf26ec-8a64-472c-86a5-918b5170ce61-audit-dir\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.580468 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.581443 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.581448 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.582140 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-router-certs\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.586183 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.586818 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/96cf26ec-8a64-472c-86a5-918b5170ce61-v4-0-config-system-session\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.599545 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fgzq\" (UniqueName: \"kubernetes.io/projected/96cf26ec-8a64-472c-86a5-918b5170ce61-kube-api-access-5fgzq\") pod \"oauth-openshift-5db8794bf8-j99ff\" (UID: \"96cf26ec-8a64-472c-86a5-918b5170ce61\") " pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.624531 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.838372 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5db8794bf8-j99ff"] Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.857255 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.892645 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 11 08:41:48 crc kubenswrapper[4629]: I1211 08:41:48.984412 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.091907 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.212938 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.213677 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.300963 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.334443 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.379007 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" event={"ID":"96cf26ec-8a64-472c-86a5-918b5170ce61","Type":"ContainerStarted","Data":"5c071a830db76507d2412b34c93bbec5a311c4066e2fdf415cc293d9a56e0798"} Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.379049 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" event={"ID":"96cf26ec-8a64-472c-86a5-918b5170ce61","Type":"ContainerStarted","Data":"2f91dff7fc4af07110187a430839661f645de9b29e0e7699b0c1a4dad6d435d4"} Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.379279 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.401204 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" podStartSLOduration=50.401185021 podStartE2EDuration="50.401185021s" podCreationTimestamp="2025-12-11 08:40:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:41:49.397913545 +0000 UTC m=+297.463329163" watchObservedRunningTime="2025-12-11 08:41:49.401185021 +0000 UTC m=+297.466600629" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.446543 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.462329 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.507982 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.550532 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.573348 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.578635 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5db8794bf8-j99ff" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.640364 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.676635 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.684034 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.744146 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 11 08:41:49 crc kubenswrapper[4629]: I1211 08:41:49.747914 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 11 08:41:50 crc kubenswrapper[4629]: I1211 08:41:50.051417 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 11 08:41:50 crc kubenswrapper[4629]: I1211 08:41:50.172156 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 11 08:41:50 crc kubenswrapper[4629]: I1211 08:41:50.198592 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 11 08:41:50 crc kubenswrapper[4629]: I1211 08:41:50.252102 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 11 08:41:50 crc kubenswrapper[4629]: I1211 08:41:50.459974 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 11 08:41:50 crc kubenswrapper[4629]: I1211 08:41:50.545739 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 11 08:41:50 crc kubenswrapper[4629]: I1211 08:41:50.727394 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 11 08:41:50 crc kubenswrapper[4629]: I1211 08:41:50.856514 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 11 08:41:50 crc kubenswrapper[4629]: I1211 08:41:50.970159 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.111507 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.129267 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.247083 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.330913 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.330984 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.397731 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.397785 4629 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="14cd17fa30381fcce30a71be2280881a94bb1e28df0b235f4ddc4af5c89acd12" exitCode=137 Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.397885 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.397944 4629 scope.go:117] "RemoveContainer" containerID="14cd17fa30381fcce30a71be2280881a94bb1e28df0b235f4ddc4af5c89acd12" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.405833 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.405889 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.405923 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.405935 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.405972 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.405994 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.406090 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.406318 4629 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.406332 4629 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.406385 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.406422 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.408105 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.413081 4629 scope.go:117] "RemoveContainer" containerID="14cd17fa30381fcce30a71be2280881a94bb1e28df0b235f4ddc4af5c89acd12" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.413340 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:41:51 crc kubenswrapper[4629]: E1211 08:41:51.414213 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14cd17fa30381fcce30a71be2280881a94bb1e28df0b235f4ddc4af5c89acd12\": container with ID starting with 14cd17fa30381fcce30a71be2280881a94bb1e28df0b235f4ddc4af5c89acd12 not found: ID does not exist" containerID="14cd17fa30381fcce30a71be2280881a94bb1e28df0b235f4ddc4af5c89acd12" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.414253 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14cd17fa30381fcce30a71be2280881a94bb1e28df0b235f4ddc4af5c89acd12"} err="failed to get container status \"14cd17fa30381fcce30a71be2280881a94bb1e28df0b235f4ddc4af5c89acd12\": rpc error: code = NotFound desc = could not find container \"14cd17fa30381fcce30a71be2280881a94bb1e28df0b235f4ddc4af5c89acd12\": container with ID starting with 14cd17fa30381fcce30a71be2280881a94bb1e28df0b235f4ddc4af5c89acd12 not found: ID does not exist" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.502507 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.507829 4629 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.507876 4629 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.507890 4629 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.565894 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 11 08:41:51 crc kubenswrapper[4629]: I1211 08:41:51.961261 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 11 08:41:52 crc kubenswrapper[4629]: I1211 08:41:52.010173 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 11 08:41:52 crc kubenswrapper[4629]: I1211 08:41:52.072451 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 11 08:41:52 crc kubenswrapper[4629]: I1211 08:41:52.209625 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 11 08:41:52 crc kubenswrapper[4629]: I1211 08:41:52.210753 4629 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 11 08:41:52 crc kubenswrapper[4629]: I1211 08:41:52.302541 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 11 08:41:52 crc kubenswrapper[4629]: I1211 08:41:52.303113 4629 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="69350956-7677-4ee0-9002-b7a463b70888" Dec 11 08:41:52 crc kubenswrapper[4629]: I1211 08:41:52.307781 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 11 08:41:52 crc kubenswrapper[4629]: I1211 08:41:52.307893 4629 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="69350956-7677-4ee0-9002-b7a463b70888" Dec 11 08:41:52 crc kubenswrapper[4629]: I1211 08:41:52.461922 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 11 08:41:52 crc kubenswrapper[4629]: I1211 08:41:52.465964 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 11 08:41:53 crc kubenswrapper[4629]: I1211 08:41:53.713825 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.605012 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n2r4l"] Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.606979 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n2r4l" podUID="ccfb1548-f379-4fb3-ba39-b138585900b4" containerName="registry-server" containerID="cri-o://0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa" gracePeriod=30 Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.610246 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8f998"] Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.610513 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8f998" podUID="c60b1455-35ba-47a0-a4d8-016c651e8c22" containerName="registry-server" containerID="cri-o://38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f" gracePeriod=30 Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.624775 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rzqjs"] Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.625018 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" podUID="128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" containerName="marketplace-operator" containerID="cri-o://26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed" gracePeriod=30 Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.629586 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8vkr"] Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.629815 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q8vkr" podUID="28a0959a-1692-4c2e-83c3-54627cd1d53a" containerName="registry-server" containerID="cri-o://bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420" gracePeriod=30 Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.645414 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-69vgp"] Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.645700 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-69vgp" podUID="ecaf1148-6695-4cdf-9ffc-06cab2716241" containerName="registry-server" containerID="cri-o://59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae" gracePeriod=30 Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.658648 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4cvmx"] Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.659411 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.678930 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4cvmx"] Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.746177 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg5f5\" (UniqueName: \"kubernetes.io/projected/9b48c9ef-aca6-40cf-b21c-58f804639e0e-kube-api-access-wg5f5\") pod \"marketplace-operator-79b997595-4cvmx\" (UID: \"9b48c9ef-aca6-40cf-b21c-58f804639e0e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.746241 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b48c9ef-aca6-40cf-b21c-58f804639e0e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4cvmx\" (UID: \"9b48c9ef-aca6-40cf-b21c-58f804639e0e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.746264 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9b48c9ef-aca6-40cf-b21c-58f804639e0e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4cvmx\" (UID: \"9b48c9ef-aca6-40cf-b21c-58f804639e0e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.846977 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg5f5\" (UniqueName: \"kubernetes.io/projected/9b48c9ef-aca6-40cf-b21c-58f804639e0e-kube-api-access-wg5f5\") pod \"marketplace-operator-79b997595-4cvmx\" (UID: \"9b48c9ef-aca6-40cf-b21c-58f804639e0e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.847028 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b48c9ef-aca6-40cf-b21c-58f804639e0e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4cvmx\" (UID: \"9b48c9ef-aca6-40cf-b21c-58f804639e0e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.847050 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9b48c9ef-aca6-40cf-b21c-58f804639e0e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4cvmx\" (UID: \"9b48c9ef-aca6-40cf-b21c-58f804639e0e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.853957 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b48c9ef-aca6-40cf-b21c-58f804639e0e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4cvmx\" (UID: \"9b48c9ef-aca6-40cf-b21c-58f804639e0e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.880750 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9b48c9ef-aca6-40cf-b21c-58f804639e0e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4cvmx\" (UID: \"9b48c9ef-aca6-40cf-b21c-58f804639e0e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.889708 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg5f5\" (UniqueName: \"kubernetes.io/projected/9b48c9ef-aca6-40cf-b21c-58f804639e0e-kube-api-access-wg5f5\") pod \"marketplace-operator-79b997595-4cvmx\" (UID: \"9b48c9ef-aca6-40cf-b21c-58f804639e0e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:07 crc kubenswrapper[4629]: I1211 08:42:07.981066 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.066751 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.152598 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfb1548-f379-4fb3-ba39-b138585900b4-utilities\") pod \"ccfb1548-f379-4fb3-ba39-b138585900b4\" (UID: \"ccfb1548-f379-4fb3-ba39-b138585900b4\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.152666 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q47jn\" (UniqueName: \"kubernetes.io/projected/ccfb1548-f379-4fb3-ba39-b138585900b4-kube-api-access-q47jn\") pod \"ccfb1548-f379-4fb3-ba39-b138585900b4\" (UID: \"ccfb1548-f379-4fb3-ba39-b138585900b4\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.152706 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfb1548-f379-4fb3-ba39-b138585900b4-catalog-content\") pod \"ccfb1548-f379-4fb3-ba39-b138585900b4\" (UID: \"ccfb1548-f379-4fb3-ba39-b138585900b4\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.154669 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccfb1548-f379-4fb3-ba39-b138585900b4-utilities" (OuterVolumeSpecName: "utilities") pod "ccfb1548-f379-4fb3-ba39-b138585900b4" (UID: "ccfb1548-f379-4fb3-ba39-b138585900b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.159100 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccfb1548-f379-4fb3-ba39-b138585900b4-kube-api-access-q47jn" (OuterVolumeSpecName: "kube-api-access-q47jn") pod "ccfb1548-f379-4fb3-ba39-b138585900b4" (UID: "ccfb1548-f379-4fb3-ba39-b138585900b4"). InnerVolumeSpecName "kube-api-access-q47jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.162697 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.169058 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.170936 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8f998" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.178193 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.224389 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccfb1548-f379-4fb3-ba39-b138585900b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccfb1548-f379-4fb3-ba39-b138585900b4" (UID: "ccfb1548-f379-4fb3-ba39-b138585900b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.253688 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecaf1148-6695-4cdf-9ffc-06cab2716241-catalog-content\") pod \"ecaf1148-6695-4cdf-9ffc-06cab2716241\" (UID: \"ecaf1148-6695-4cdf-9ffc-06cab2716241\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.254047 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kbnb\" (UniqueName: \"kubernetes.io/projected/28a0959a-1692-4c2e-83c3-54627cd1d53a-kube-api-access-4kbnb\") pod \"28a0959a-1692-4c2e-83c3-54627cd1d53a\" (UID: \"28a0959a-1692-4c2e-83c3-54627cd1d53a\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.254149 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-marketplace-operator-metrics\") pod \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\" (UID: \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.254249 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm8t6\" (UniqueName: \"kubernetes.io/projected/c60b1455-35ba-47a0-a4d8-016c651e8c22-kube-api-access-tm8t6\") pod \"c60b1455-35ba-47a0-a4d8-016c651e8c22\" (UID: \"c60b1455-35ba-47a0-a4d8-016c651e8c22\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.254390 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwdfq\" (UniqueName: \"kubernetes.io/projected/ecaf1148-6695-4cdf-9ffc-06cab2716241-kube-api-access-wwdfq\") pod \"ecaf1148-6695-4cdf-9ffc-06cab2716241\" (UID: \"ecaf1148-6695-4cdf-9ffc-06cab2716241\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.254488 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28a0959a-1692-4c2e-83c3-54627cd1d53a-catalog-content\") pod \"28a0959a-1692-4c2e-83c3-54627cd1d53a\" (UID: \"28a0959a-1692-4c2e-83c3-54627cd1d53a\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.254603 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c60b1455-35ba-47a0-a4d8-016c651e8c22-catalog-content\") pod \"c60b1455-35ba-47a0-a4d8-016c651e8c22\" (UID: \"c60b1455-35ba-47a0-a4d8-016c651e8c22\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.254694 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28a0959a-1692-4c2e-83c3-54627cd1d53a-utilities\") pod \"28a0959a-1692-4c2e-83c3-54627cd1d53a\" (UID: \"28a0959a-1692-4c2e-83c3-54627cd1d53a\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.254814 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-marketplace-trusted-ca\") pod \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\" (UID: \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.254945 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c60b1455-35ba-47a0-a4d8-016c651e8c22-utilities\") pod \"c60b1455-35ba-47a0-a4d8-016c651e8c22\" (UID: \"c60b1455-35ba-47a0-a4d8-016c651e8c22\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.255053 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xv9cf\" (UniqueName: \"kubernetes.io/projected/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-kube-api-access-xv9cf\") pod \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\" (UID: \"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.255161 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecaf1148-6695-4cdf-9ffc-06cab2716241-utilities\") pod \"ecaf1148-6695-4cdf-9ffc-06cab2716241\" (UID: \"ecaf1148-6695-4cdf-9ffc-06cab2716241\") " Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.257439 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfb1548-f379-4fb3-ba39-b138585900b4-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.272028 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q47jn\" (UniqueName: \"kubernetes.io/projected/ccfb1548-f379-4fb3-ba39-b138585900b4-kube-api-access-q47jn\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.272049 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfb1548-f379-4fb3-ba39-b138585900b4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.258573 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" (UID: "128e77ce-6b4b-4c28-a10b-fa3b84b13fcc"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.259573 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecaf1148-6695-4cdf-9ffc-06cab2716241-utilities" (OuterVolumeSpecName: "utilities") pod "ecaf1148-6695-4cdf-9ffc-06cab2716241" (UID: "ecaf1148-6695-4cdf-9ffc-06cab2716241"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.259615 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c60b1455-35ba-47a0-a4d8-016c651e8c22-utilities" (OuterVolumeSpecName: "utilities") pod "c60b1455-35ba-47a0-a4d8-016c651e8c22" (UID: "c60b1455-35ba-47a0-a4d8-016c651e8c22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.262449 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" (UID: "128e77ce-6b4b-4c28-a10b-fa3b84b13fcc"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.263900 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28a0959a-1692-4c2e-83c3-54627cd1d53a-utilities" (OuterVolumeSpecName: "utilities") pod "28a0959a-1692-4c2e-83c3-54627cd1d53a" (UID: "28a0959a-1692-4c2e-83c3-54627cd1d53a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.264010 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28a0959a-1692-4c2e-83c3-54627cd1d53a-kube-api-access-4kbnb" (OuterVolumeSpecName: "kube-api-access-4kbnb") pod "28a0959a-1692-4c2e-83c3-54627cd1d53a" (UID: "28a0959a-1692-4c2e-83c3-54627cd1d53a"). InnerVolumeSpecName "kube-api-access-4kbnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.266088 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecaf1148-6695-4cdf-9ffc-06cab2716241-kube-api-access-wwdfq" (OuterVolumeSpecName: "kube-api-access-wwdfq") pod "ecaf1148-6695-4cdf-9ffc-06cab2716241" (UID: "ecaf1148-6695-4cdf-9ffc-06cab2716241"). InnerVolumeSpecName "kube-api-access-wwdfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.271209 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c60b1455-35ba-47a0-a4d8-016c651e8c22-kube-api-access-tm8t6" (OuterVolumeSpecName: "kube-api-access-tm8t6") pod "c60b1455-35ba-47a0-a4d8-016c651e8c22" (UID: "c60b1455-35ba-47a0-a4d8-016c651e8c22"). InnerVolumeSpecName "kube-api-access-tm8t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.271513 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-kube-api-access-xv9cf" (OuterVolumeSpecName: "kube-api-access-xv9cf") pod "128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" (UID: "128e77ce-6b4b-4c28-a10b-fa3b84b13fcc"). InnerVolumeSpecName "kube-api-access-xv9cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.290506 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28a0959a-1692-4c2e-83c3-54627cd1d53a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28a0959a-1692-4c2e-83c3-54627cd1d53a" (UID: "28a0959a-1692-4c2e-83c3-54627cd1d53a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.322365 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c60b1455-35ba-47a0-a4d8-016c651e8c22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c60b1455-35ba-47a0-a4d8-016c651e8c22" (UID: "c60b1455-35ba-47a0-a4d8-016c651e8c22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.373441 4629 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.373490 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c60b1455-35ba-47a0-a4d8-016c651e8c22-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.373506 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xv9cf\" (UniqueName: \"kubernetes.io/projected/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-kube-api-access-xv9cf\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.373519 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecaf1148-6695-4cdf-9ffc-06cab2716241-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.373532 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kbnb\" (UniqueName: \"kubernetes.io/projected/28a0959a-1692-4c2e-83c3-54627cd1d53a-kube-api-access-4kbnb\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.373543 4629 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.373555 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm8t6\" (UniqueName: \"kubernetes.io/projected/c60b1455-35ba-47a0-a4d8-016c651e8c22-kube-api-access-tm8t6\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.373565 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwdfq\" (UniqueName: \"kubernetes.io/projected/ecaf1148-6695-4cdf-9ffc-06cab2716241-kube-api-access-wwdfq\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.373623 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28a0959a-1692-4c2e-83c3-54627cd1d53a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.373635 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c60b1455-35ba-47a0-a4d8-016c651e8c22-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.373643 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28a0959a-1692-4c2e-83c3-54627cd1d53a-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.386517 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecaf1148-6695-4cdf-9ffc-06cab2716241-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ecaf1148-6695-4cdf-9ffc-06cab2716241" (UID: "ecaf1148-6695-4cdf-9ffc-06cab2716241"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.475110 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecaf1148-6695-4cdf-9ffc-06cab2716241-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.496513 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4cvmx"] Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.625576 4629 generic.go:334] "Generic (PLEG): container finished" podID="ccfb1548-f379-4fb3-ba39-b138585900b4" containerID="0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa" exitCode=0 Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.625691 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2r4l" event={"ID":"ccfb1548-f379-4fb3-ba39-b138585900b4","Type":"ContainerDied","Data":"0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa"} Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.625735 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2r4l" event={"ID":"ccfb1548-f379-4fb3-ba39-b138585900b4","Type":"ContainerDied","Data":"79acdfe15c099e03d4d026e4952cf15f081e5b8925575bff8652260494034bbf"} Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.625754 4629 scope.go:117] "RemoveContainer" containerID="0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.625892 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2r4l" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.632398 4629 generic.go:334] "Generic (PLEG): container finished" podID="28a0959a-1692-4c2e-83c3-54627cd1d53a" containerID="bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420" exitCode=0 Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.632519 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8vkr" event={"ID":"28a0959a-1692-4c2e-83c3-54627cd1d53a","Type":"ContainerDied","Data":"bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420"} Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.632572 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8vkr" event={"ID":"28a0959a-1692-4c2e-83c3-54627cd1d53a","Type":"ContainerDied","Data":"83e827b82abeabe17d0310eb1d1e0fe1939bf0a065fb3dfd5b71b3d7b2949662"} Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.632595 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8vkr" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.642326 4629 generic.go:334] "Generic (PLEG): container finished" podID="c60b1455-35ba-47a0-a4d8-016c651e8c22" containerID="38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f" exitCode=0 Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.642421 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8f998" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.642459 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8f998" event={"ID":"c60b1455-35ba-47a0-a4d8-016c651e8c22","Type":"ContainerDied","Data":"38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f"} Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.642526 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8f998" event={"ID":"c60b1455-35ba-47a0-a4d8-016c651e8c22","Type":"ContainerDied","Data":"4db31d63398f2d468edb5ed52bab9df25d1ff358c21f88da55c3aa0cd6d360d8"} Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.649171 4629 scope.go:117] "RemoveContainer" containerID="373d7b045bc8fae5fb716e3f138c163ef5b68f2199a9952045f1fb242c1aa0c7" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.657886 4629 generic.go:334] "Generic (PLEG): container finished" podID="ecaf1148-6695-4cdf-9ffc-06cab2716241" containerID="59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae" exitCode=0 Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.658037 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69vgp" event={"ID":"ecaf1148-6695-4cdf-9ffc-06cab2716241","Type":"ContainerDied","Data":"59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae"} Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.658092 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69vgp" event={"ID":"ecaf1148-6695-4cdf-9ffc-06cab2716241","Type":"ContainerDied","Data":"461194cbec6e758e224db7142633f02e346cefa646c63477ccbf6d0390d2b302"} Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.658204 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69vgp" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.661918 4629 generic.go:334] "Generic (PLEG): container finished" podID="128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" containerID="26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed" exitCode=0 Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.662169 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" event={"ID":"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc","Type":"ContainerDied","Data":"26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed"} Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.662240 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" event={"ID":"128e77ce-6b4b-4c28-a10b-fa3b84b13fcc","Type":"ContainerDied","Data":"39e0d635f9c4c3f1a4f983b4b537ef0a1f925ec3e32c1602829456e2f268ae56"} Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.662365 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rzqjs" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.678675 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" event={"ID":"9b48c9ef-aca6-40cf-b21c-58f804639e0e","Type":"ContainerStarted","Data":"1eda4f1bbe1728d607e0127c8f1db7569af7dccab7032eebeec93366326c38c5"} Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.686590 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n2r4l"] Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.691835 4629 scope.go:117] "RemoveContainer" containerID="4874bac62fb121e930fb70bf5fb575aff02b20d49f2dad3b77f63ce41896eef6" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.694924 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n2r4l"] Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.707359 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8vkr"] Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.713204 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8vkr"] Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.722998 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8f998"] Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.729350 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8f998"] Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.730621 4629 scope.go:117] "RemoveContainer" containerID="0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.731174 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa\": container with ID starting with 0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa not found: ID does not exist" containerID="0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.731206 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa"} err="failed to get container status \"0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa\": rpc error: code = NotFound desc = could not find container \"0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa\": container with ID starting with 0a931b7d0e5f15983d174638d5032d411993ae6c5cab61fecec8ce80032497fa not found: ID does not exist" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.731230 4629 scope.go:117] "RemoveContainer" containerID="373d7b045bc8fae5fb716e3f138c163ef5b68f2199a9952045f1fb242c1aa0c7" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.731554 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"373d7b045bc8fae5fb716e3f138c163ef5b68f2199a9952045f1fb242c1aa0c7\": container with ID starting with 373d7b045bc8fae5fb716e3f138c163ef5b68f2199a9952045f1fb242c1aa0c7 not found: ID does not exist" containerID="373d7b045bc8fae5fb716e3f138c163ef5b68f2199a9952045f1fb242c1aa0c7" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.731574 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"373d7b045bc8fae5fb716e3f138c163ef5b68f2199a9952045f1fb242c1aa0c7"} err="failed to get container status \"373d7b045bc8fae5fb716e3f138c163ef5b68f2199a9952045f1fb242c1aa0c7\": rpc error: code = NotFound desc = could not find container \"373d7b045bc8fae5fb716e3f138c163ef5b68f2199a9952045f1fb242c1aa0c7\": container with ID starting with 373d7b045bc8fae5fb716e3f138c163ef5b68f2199a9952045f1fb242c1aa0c7 not found: ID does not exist" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.731588 4629 scope.go:117] "RemoveContainer" containerID="4874bac62fb121e930fb70bf5fb575aff02b20d49f2dad3b77f63ce41896eef6" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.731900 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4874bac62fb121e930fb70bf5fb575aff02b20d49f2dad3b77f63ce41896eef6\": container with ID starting with 4874bac62fb121e930fb70bf5fb575aff02b20d49f2dad3b77f63ce41896eef6 not found: ID does not exist" containerID="4874bac62fb121e930fb70bf5fb575aff02b20d49f2dad3b77f63ce41896eef6" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.731920 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4874bac62fb121e930fb70bf5fb575aff02b20d49f2dad3b77f63ce41896eef6"} err="failed to get container status \"4874bac62fb121e930fb70bf5fb575aff02b20d49f2dad3b77f63ce41896eef6\": rpc error: code = NotFound desc = could not find container \"4874bac62fb121e930fb70bf5fb575aff02b20d49f2dad3b77f63ce41896eef6\": container with ID starting with 4874bac62fb121e930fb70bf5fb575aff02b20d49f2dad3b77f63ce41896eef6 not found: ID does not exist" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.731933 4629 scope.go:117] "RemoveContainer" containerID="bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.741855 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rzqjs"] Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.745727 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rzqjs"] Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.754157 4629 scope.go:117] "RemoveContainer" containerID="e657d64611c3d908b9654ed4e7933bcef9fe15a08cc162f71077d398bb64d400" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.758186 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-69vgp"] Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.763630 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-69vgp"] Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.818128 4629 scope.go:117] "RemoveContainer" containerID="91023925122b1e130dd383ad68b646646d822b11bf66a2a5bf1c401966d0e34d" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.832914 4629 scope.go:117] "RemoveContainer" containerID="bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.833916 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420\": container with ID starting with bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420 not found: ID does not exist" containerID="bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.833989 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420"} err="failed to get container status \"bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420\": rpc error: code = NotFound desc = could not find container \"bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420\": container with ID starting with bdbe606909907de8435c826e2ff991015108b66399c062c71ca71cd282b45420 not found: ID does not exist" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.834028 4629 scope.go:117] "RemoveContainer" containerID="e657d64611c3d908b9654ed4e7933bcef9fe15a08cc162f71077d398bb64d400" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.834782 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e657d64611c3d908b9654ed4e7933bcef9fe15a08cc162f71077d398bb64d400\": container with ID starting with e657d64611c3d908b9654ed4e7933bcef9fe15a08cc162f71077d398bb64d400 not found: ID does not exist" containerID="e657d64611c3d908b9654ed4e7933bcef9fe15a08cc162f71077d398bb64d400" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.834818 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e657d64611c3d908b9654ed4e7933bcef9fe15a08cc162f71077d398bb64d400"} err="failed to get container status \"e657d64611c3d908b9654ed4e7933bcef9fe15a08cc162f71077d398bb64d400\": rpc error: code = NotFound desc = could not find container \"e657d64611c3d908b9654ed4e7933bcef9fe15a08cc162f71077d398bb64d400\": container with ID starting with e657d64611c3d908b9654ed4e7933bcef9fe15a08cc162f71077d398bb64d400 not found: ID does not exist" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.834835 4629 scope.go:117] "RemoveContainer" containerID="91023925122b1e130dd383ad68b646646d822b11bf66a2a5bf1c401966d0e34d" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.835365 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91023925122b1e130dd383ad68b646646d822b11bf66a2a5bf1c401966d0e34d\": container with ID starting with 91023925122b1e130dd383ad68b646646d822b11bf66a2a5bf1c401966d0e34d not found: ID does not exist" containerID="91023925122b1e130dd383ad68b646646d822b11bf66a2a5bf1c401966d0e34d" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.835406 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91023925122b1e130dd383ad68b646646d822b11bf66a2a5bf1c401966d0e34d"} err="failed to get container status \"91023925122b1e130dd383ad68b646646d822b11bf66a2a5bf1c401966d0e34d\": rpc error: code = NotFound desc = could not find container \"91023925122b1e130dd383ad68b646646d822b11bf66a2a5bf1c401966d0e34d\": container with ID starting with 91023925122b1e130dd383ad68b646646d822b11bf66a2a5bf1c401966d0e34d not found: ID does not exist" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.835423 4629 scope.go:117] "RemoveContainer" containerID="38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.852439 4629 scope.go:117] "RemoveContainer" containerID="f29b61cfc37fa66bffae8b0f9908855b925c5a6acb348820cb704096cd53daed" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.877739 4629 scope.go:117] "RemoveContainer" containerID="7cf74e82ff2a57d0552522b860fbbe85844e94d007f6da653dec65d6187369e1" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.897165 4629 scope.go:117] "RemoveContainer" containerID="38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.897623 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f\": container with ID starting with 38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f not found: ID does not exist" containerID="38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.897653 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f"} err="failed to get container status \"38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f\": rpc error: code = NotFound desc = could not find container \"38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f\": container with ID starting with 38e991bfbac191409b02b393ceef1196565066d4f4a15f954fe28c92b3a9ce2f not found: ID does not exist" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.897700 4629 scope.go:117] "RemoveContainer" containerID="f29b61cfc37fa66bffae8b0f9908855b925c5a6acb348820cb704096cd53daed" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.898083 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f29b61cfc37fa66bffae8b0f9908855b925c5a6acb348820cb704096cd53daed\": container with ID starting with f29b61cfc37fa66bffae8b0f9908855b925c5a6acb348820cb704096cd53daed not found: ID does not exist" containerID="f29b61cfc37fa66bffae8b0f9908855b925c5a6acb348820cb704096cd53daed" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.898164 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f29b61cfc37fa66bffae8b0f9908855b925c5a6acb348820cb704096cd53daed"} err="failed to get container status \"f29b61cfc37fa66bffae8b0f9908855b925c5a6acb348820cb704096cd53daed\": rpc error: code = NotFound desc = could not find container \"f29b61cfc37fa66bffae8b0f9908855b925c5a6acb348820cb704096cd53daed\": container with ID starting with f29b61cfc37fa66bffae8b0f9908855b925c5a6acb348820cb704096cd53daed not found: ID does not exist" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.898223 4629 scope.go:117] "RemoveContainer" containerID="7cf74e82ff2a57d0552522b860fbbe85844e94d007f6da653dec65d6187369e1" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.898658 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cf74e82ff2a57d0552522b860fbbe85844e94d007f6da653dec65d6187369e1\": container with ID starting with 7cf74e82ff2a57d0552522b860fbbe85844e94d007f6da653dec65d6187369e1 not found: ID does not exist" containerID="7cf74e82ff2a57d0552522b860fbbe85844e94d007f6da653dec65d6187369e1" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.898689 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cf74e82ff2a57d0552522b860fbbe85844e94d007f6da653dec65d6187369e1"} err="failed to get container status \"7cf74e82ff2a57d0552522b860fbbe85844e94d007f6da653dec65d6187369e1\": rpc error: code = NotFound desc = could not find container \"7cf74e82ff2a57d0552522b860fbbe85844e94d007f6da653dec65d6187369e1\": container with ID starting with 7cf74e82ff2a57d0552522b860fbbe85844e94d007f6da653dec65d6187369e1 not found: ID does not exist" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.898727 4629 scope.go:117] "RemoveContainer" containerID="59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.922066 4629 scope.go:117] "RemoveContainer" containerID="f3b6f6792b49d9d807f7359fbfb32c85663c623b9032a863acb3a00aef93be47" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.938669 4629 scope.go:117] "RemoveContainer" containerID="d105d0b909583f816b453856eb5e3964ac57b8eaaad00fd43f86314c02c87e7e" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.954335 4629 scope.go:117] "RemoveContainer" containerID="59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.955355 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae\": container with ID starting with 59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae not found: ID does not exist" containerID="59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.955418 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae"} err="failed to get container status \"59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae\": rpc error: code = NotFound desc = could not find container \"59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae\": container with ID starting with 59aafd9ab918ec6b4c8949bc924ae746f613f24a6c5b4c73900378dde5c755ae not found: ID does not exist" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.955476 4629 scope.go:117] "RemoveContainer" containerID="f3b6f6792b49d9d807f7359fbfb32c85663c623b9032a863acb3a00aef93be47" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.955894 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3b6f6792b49d9d807f7359fbfb32c85663c623b9032a863acb3a00aef93be47\": container with ID starting with f3b6f6792b49d9d807f7359fbfb32c85663c623b9032a863acb3a00aef93be47 not found: ID does not exist" containerID="f3b6f6792b49d9d807f7359fbfb32c85663c623b9032a863acb3a00aef93be47" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.955958 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3b6f6792b49d9d807f7359fbfb32c85663c623b9032a863acb3a00aef93be47"} err="failed to get container status \"f3b6f6792b49d9d807f7359fbfb32c85663c623b9032a863acb3a00aef93be47\": rpc error: code = NotFound desc = could not find container \"f3b6f6792b49d9d807f7359fbfb32c85663c623b9032a863acb3a00aef93be47\": container with ID starting with f3b6f6792b49d9d807f7359fbfb32c85663c623b9032a863acb3a00aef93be47 not found: ID does not exist" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.956024 4629 scope.go:117] "RemoveContainer" containerID="d105d0b909583f816b453856eb5e3964ac57b8eaaad00fd43f86314c02c87e7e" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.956378 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d105d0b909583f816b453856eb5e3964ac57b8eaaad00fd43f86314c02c87e7e\": container with ID starting with d105d0b909583f816b453856eb5e3964ac57b8eaaad00fd43f86314c02c87e7e not found: ID does not exist" containerID="d105d0b909583f816b453856eb5e3964ac57b8eaaad00fd43f86314c02c87e7e" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.956431 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d105d0b909583f816b453856eb5e3964ac57b8eaaad00fd43f86314c02c87e7e"} err="failed to get container status \"d105d0b909583f816b453856eb5e3964ac57b8eaaad00fd43f86314c02c87e7e\": rpc error: code = NotFound desc = could not find container \"d105d0b909583f816b453856eb5e3964ac57b8eaaad00fd43f86314c02c87e7e\": container with ID starting with d105d0b909583f816b453856eb5e3964ac57b8eaaad00fd43f86314c02c87e7e not found: ID does not exist" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.956468 4629 scope.go:117] "RemoveContainer" containerID="26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.974122 4629 scope.go:117] "RemoveContainer" containerID="26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed" Dec 11 08:42:08 crc kubenswrapper[4629]: E1211 08:42:08.975300 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed\": container with ID starting with 26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed not found: ID does not exist" containerID="26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed" Dec 11 08:42:08 crc kubenswrapper[4629]: I1211 08:42:08.975338 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed"} err="failed to get container status \"26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed\": rpc error: code = NotFound desc = could not find container \"26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed\": container with ID starting with 26f968a27a80a3384e8fda2e065685702b373c1995d25ab782ecaeaa4f92d8ed not found: ID does not exist" Dec 11 08:42:09 crc kubenswrapper[4629]: I1211 08:42:09.687136 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" event={"ID":"9b48c9ef-aca6-40cf-b21c-58f804639e0e","Type":"ContainerStarted","Data":"5e34bdb88bc4b57aaaff67ba5a2aed91e89a5eb7363fe8e28bb7889e365ad39d"} Dec 11 08:42:09 crc kubenswrapper[4629]: I1211 08:42:09.688693 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:09 crc kubenswrapper[4629]: I1211 08:42:09.699001 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" Dec 11 08:42:09 crc kubenswrapper[4629]: I1211 08:42:09.717566 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4cvmx" podStartSLOduration=2.7175079589999998 podStartE2EDuration="2.717507959s" podCreationTimestamp="2025-12-11 08:42:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:42:09.701624838 +0000 UTC m=+317.767040466" watchObservedRunningTime="2025-12-11 08:42:09.717507959 +0000 UTC m=+317.782923567" Dec 11 08:42:10 crc kubenswrapper[4629]: I1211 08:42:10.206294 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" path="/var/lib/kubelet/pods/128e77ce-6b4b-4c28-a10b-fa3b84b13fcc/volumes" Dec 11 08:42:10 crc kubenswrapper[4629]: I1211 08:42:10.207274 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28a0959a-1692-4c2e-83c3-54627cd1d53a" path="/var/lib/kubelet/pods/28a0959a-1692-4c2e-83c3-54627cd1d53a/volumes" Dec 11 08:42:10 crc kubenswrapper[4629]: I1211 08:42:10.208207 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c60b1455-35ba-47a0-a4d8-016c651e8c22" path="/var/lib/kubelet/pods/c60b1455-35ba-47a0-a4d8-016c651e8c22/volumes" Dec 11 08:42:10 crc kubenswrapper[4629]: I1211 08:42:10.209650 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccfb1548-f379-4fb3-ba39-b138585900b4" path="/var/lib/kubelet/pods/ccfb1548-f379-4fb3-ba39-b138585900b4/volumes" Dec 11 08:42:10 crc kubenswrapper[4629]: I1211 08:42:10.210488 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecaf1148-6695-4cdf-9ffc-06cab2716241" path="/var/lib/kubelet/pods/ecaf1148-6695-4cdf-9ffc-06cab2716241/volumes" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.204381 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2c7tp"] Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.205239 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" podUID="dbb99da1-1049-4e03-bc19-49988fc37947" containerName="controller-manager" containerID="cri-o://f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67" gracePeriod=30 Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.314688 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v"] Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.314948 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" podUID="ec0bf78c-4e35-4129-8976-c8d805991843" containerName="route-controller-manager" containerID="cri-o://ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2" gracePeriod=30 Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.555039 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.654063 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.730087 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-proxy-ca-bundles\") pod \"dbb99da1-1049-4e03-bc19-49988fc37947\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.730187 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wvff\" (UniqueName: \"kubernetes.io/projected/dbb99da1-1049-4e03-bc19-49988fc37947-kube-api-access-6wvff\") pod \"dbb99da1-1049-4e03-bc19-49988fc37947\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.730222 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbb99da1-1049-4e03-bc19-49988fc37947-serving-cert\") pod \"dbb99da1-1049-4e03-bc19-49988fc37947\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.730250 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-config\") pod \"dbb99da1-1049-4e03-bc19-49988fc37947\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.730299 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-client-ca\") pod \"dbb99da1-1049-4e03-bc19-49988fc37947\" (UID: \"dbb99da1-1049-4e03-bc19-49988fc37947\") " Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.731315 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-config" (OuterVolumeSpecName: "config") pod "dbb99da1-1049-4e03-bc19-49988fc37947" (UID: "dbb99da1-1049-4e03-bc19-49988fc37947"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.731367 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-client-ca" (OuterVolumeSpecName: "client-ca") pod "dbb99da1-1049-4e03-bc19-49988fc37947" (UID: "dbb99da1-1049-4e03-bc19-49988fc37947"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.731412 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "dbb99da1-1049-4e03-bc19-49988fc37947" (UID: "dbb99da1-1049-4e03-bc19-49988fc37947"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.735447 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbb99da1-1049-4e03-bc19-49988fc37947-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dbb99da1-1049-4e03-bc19-49988fc37947" (UID: "dbb99da1-1049-4e03-bc19-49988fc37947"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.736015 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbb99da1-1049-4e03-bc19-49988fc37947-kube-api-access-6wvff" (OuterVolumeSpecName: "kube-api-access-6wvff") pod "dbb99da1-1049-4e03-bc19-49988fc37947" (UID: "dbb99da1-1049-4e03-bc19-49988fc37947"). InnerVolumeSpecName "kube-api-access-6wvff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.831858 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ec0bf78c-4e35-4129-8976-c8d805991843-client-ca\") pod \"ec0bf78c-4e35-4129-8976-c8d805991843\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.832380 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec0bf78c-4e35-4129-8976-c8d805991843-config\") pod \"ec0bf78c-4e35-4129-8976-c8d805991843\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.832429 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9q9w\" (UniqueName: \"kubernetes.io/projected/ec0bf78c-4e35-4129-8976-c8d805991843-kube-api-access-g9q9w\") pod \"ec0bf78c-4e35-4129-8976-c8d805991843\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.832501 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec0bf78c-4e35-4129-8976-c8d805991843-serving-cert\") pod \"ec0bf78c-4e35-4129-8976-c8d805991843\" (UID: \"ec0bf78c-4e35-4129-8976-c8d805991843\") " Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.832750 4629 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.832766 4629 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.832777 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wvff\" (UniqueName: \"kubernetes.io/projected/dbb99da1-1049-4e03-bc19-49988fc37947-kube-api-access-6wvff\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.832787 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbb99da1-1049-4e03-bc19-49988fc37947-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.832796 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbb99da1-1049-4e03-bc19-49988fc37947-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.833366 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec0bf78c-4e35-4129-8976-c8d805991843-config" (OuterVolumeSpecName: "config") pod "ec0bf78c-4e35-4129-8976-c8d805991843" (UID: "ec0bf78c-4e35-4129-8976-c8d805991843"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.833501 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec0bf78c-4e35-4129-8976-c8d805991843-client-ca" (OuterVolumeSpecName: "client-ca") pod "ec0bf78c-4e35-4129-8976-c8d805991843" (UID: "ec0bf78c-4e35-4129-8976-c8d805991843"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.836673 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec0bf78c-4e35-4129-8976-c8d805991843-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ec0bf78c-4e35-4129-8976-c8d805991843" (UID: "ec0bf78c-4e35-4129-8976-c8d805991843"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.836715 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec0bf78c-4e35-4129-8976-c8d805991843-kube-api-access-g9q9w" (OuterVolumeSpecName: "kube-api-access-g9q9w") pod "ec0bf78c-4e35-4129-8976-c8d805991843" (UID: "ec0bf78c-4e35-4129-8976-c8d805991843"). InnerVolumeSpecName "kube-api-access-g9q9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.910395 4629 generic.go:334] "Generic (PLEG): container finished" podID="dbb99da1-1049-4e03-bc19-49988fc37947" containerID="f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67" exitCode=0 Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.910452 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" event={"ID":"dbb99da1-1049-4e03-bc19-49988fc37947","Type":"ContainerDied","Data":"f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67"} Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.910541 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" event={"ID":"dbb99da1-1049-4e03-bc19-49988fc37947","Type":"ContainerDied","Data":"fcdeddb96f242d5fec813d5cecdf37a59ac447ba693c97d91492a0e9c2a40e14"} Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.910562 4629 scope.go:117] "RemoveContainer" containerID="f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.910515 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2c7tp" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.912003 4629 generic.go:334] "Generic (PLEG): container finished" podID="ec0bf78c-4e35-4129-8976-c8d805991843" containerID="ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2" exitCode=0 Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.912095 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.912121 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" event={"ID":"ec0bf78c-4e35-4129-8976-c8d805991843","Type":"ContainerDied","Data":"ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2"} Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.912151 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v" event={"ID":"ec0bf78c-4e35-4129-8976-c8d805991843","Type":"ContainerDied","Data":"0052ff9cceb9ed9b182481d4b7d1dc60eb7b6265d2bed65849e1826f40b7ee0f"} Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.933663 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec0bf78c-4e35-4129-8976-c8d805991843-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.933686 4629 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ec0bf78c-4e35-4129-8976-c8d805991843-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.933696 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec0bf78c-4e35-4129-8976-c8d805991843-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.933705 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9q9w\" (UniqueName: \"kubernetes.io/projected/ec0bf78c-4e35-4129-8976-c8d805991843-kube-api-access-g9q9w\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.935988 4629 scope.go:117] "RemoveContainer" containerID="f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67" Dec 11 08:42:40 crc kubenswrapper[4629]: E1211 08:42:40.941150 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67\": container with ID starting with f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67 not found: ID does not exist" containerID="f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.942837 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67"} err="failed to get container status \"f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67\": rpc error: code = NotFound desc = could not find container \"f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67\": container with ID starting with f4077ceec0c5eaf4c4c3ad47c8bf3dca86251d3349ee1d6e852f8f25406b6d67 not found: ID does not exist" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.943001 4629 scope.go:117] "RemoveContainer" containerID="ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.943995 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v"] Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.949556 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cc5v"] Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.954670 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2c7tp"] Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.958156 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2c7tp"] Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.964784 4629 scope.go:117] "RemoveContainer" containerID="ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2" Dec 11 08:42:40 crc kubenswrapper[4629]: E1211 08:42:40.965339 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2\": container with ID starting with ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2 not found: ID does not exist" containerID="ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2" Dec 11 08:42:40 crc kubenswrapper[4629]: I1211 08:42:40.965368 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2"} err="failed to get container status \"ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2\": rpc error: code = NotFound desc = could not find container \"ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2\": container with ID starting with ac270ad3d3d8b28d5114dcd009df937c381b2cf03627fba472c7f0e9428eecd2 not found: ID does not exist" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.214540 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6787ff745b-q4747"] Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.215545 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a0959a-1692-4c2e-83c3-54627cd1d53a" containerName="extract-content" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.215671 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a0959a-1692-4c2e-83c3-54627cd1d53a" containerName="extract-content" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.215736 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0bf78c-4e35-4129-8976-c8d805991843" containerName="route-controller-manager" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.215791 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0bf78c-4e35-4129-8976-c8d805991843" containerName="route-controller-manager" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.215876 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c60b1455-35ba-47a0-a4d8-016c651e8c22" containerName="extract-content" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.215935 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c60b1455-35ba-47a0-a4d8-016c651e8c22" containerName="extract-content" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.216022 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a0959a-1692-4c2e-83c3-54627cd1d53a" containerName="registry-server" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.216084 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a0959a-1692-4c2e-83c3-54627cd1d53a" containerName="registry-server" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.216142 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccfb1548-f379-4fb3-ba39-b138585900b4" containerName="registry-server" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.216195 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccfb1548-f379-4fb3-ba39-b138585900b4" containerName="registry-server" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.216291 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c60b1455-35ba-47a0-a4d8-016c651e8c22" containerName="extract-utilities" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.216366 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c60b1455-35ba-47a0-a4d8-016c651e8c22" containerName="extract-utilities" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.216451 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecaf1148-6695-4cdf-9ffc-06cab2716241" containerName="extract-utilities" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.216559 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecaf1148-6695-4cdf-9ffc-06cab2716241" containerName="extract-utilities" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.216648 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecaf1148-6695-4cdf-9ffc-06cab2716241" containerName="registry-server" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.216731 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecaf1148-6695-4cdf-9ffc-06cab2716241" containerName="registry-server" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.216822 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c60b1455-35ba-47a0-a4d8-016c651e8c22" containerName="registry-server" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.216904 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c60b1455-35ba-47a0-a4d8-016c651e8c22" containerName="registry-server" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.216967 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a0959a-1692-4c2e-83c3-54627cd1d53a" containerName="extract-utilities" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.217024 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a0959a-1692-4c2e-83c3-54627cd1d53a" containerName="extract-utilities" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.217103 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccfb1548-f379-4fb3-ba39-b138585900b4" containerName="extract-content" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.217165 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccfb1548-f379-4fb3-ba39-b138585900b4" containerName="extract-content" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.217223 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" containerName="marketplace-operator" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.217282 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" containerName="marketplace-operator" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.217346 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbb99da1-1049-4e03-bc19-49988fc37947" containerName="controller-manager" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.217408 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbb99da1-1049-4e03-bc19-49988fc37947" containerName="controller-manager" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.217472 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecaf1148-6695-4cdf-9ffc-06cab2716241" containerName="extract-content" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.217534 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecaf1148-6695-4cdf-9ffc-06cab2716241" containerName="extract-content" Dec 11 08:42:41 crc kubenswrapper[4629]: E1211 08:42:41.217608 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccfb1548-f379-4fb3-ba39-b138585900b4" containerName="extract-utilities" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.217674 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccfb1548-f379-4fb3-ba39-b138585900b4" containerName="extract-utilities" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.217825 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccfb1548-f379-4fb3-ba39-b138585900b4" containerName="registry-server" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.217922 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbb99da1-1049-4e03-bc19-49988fc37947" containerName="controller-manager" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.217983 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="28a0959a-1692-4c2e-83c3-54627cd1d53a" containerName="registry-server" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.218061 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="128e77ce-6b4b-4c28-a10b-fa3b84b13fcc" containerName="marketplace-operator" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.218126 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecaf1148-6695-4cdf-9ffc-06cab2716241" containerName="registry-server" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.218186 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="c60b1455-35ba-47a0-a4d8-016c651e8c22" containerName="registry-server" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.218244 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec0bf78c-4e35-4129-8976-c8d805991843" containerName="route-controller-manager" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.218667 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.220829 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.221285 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh"] Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.221806 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.222167 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.223631 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.223667 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.223950 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.224828 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.225688 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.228265 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.228946 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.229291 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.229505 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.229620 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.241100 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.241465 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6787ff745b-q4747"] Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.247446 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh"] Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.340088 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a238584f-2a07-4a8c-b137-eea8509b7acf-serving-cert\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.340150 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-config\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.340184 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e07bde2a-e592-4ac1-874b-3dc4a01d061d-config\") pod \"route-controller-manager-896d4d4c6-cr4gh\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.340217 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e07bde2a-e592-4ac1-874b-3dc4a01d061d-client-ca\") pod \"route-controller-manager-896d4d4c6-cr4gh\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.340240 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c75z\" (UniqueName: \"kubernetes.io/projected/a238584f-2a07-4a8c-b137-eea8509b7acf-kube-api-access-8c75z\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.340267 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e07bde2a-e592-4ac1-874b-3dc4a01d061d-serving-cert\") pod \"route-controller-manager-896d4d4c6-cr4gh\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.340288 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94b9d\" (UniqueName: \"kubernetes.io/projected/e07bde2a-e592-4ac1-874b-3dc4a01d061d-kube-api-access-94b9d\") pod \"route-controller-manager-896d4d4c6-cr4gh\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.340308 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-client-ca\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.340326 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-proxy-ca-bundles\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.441769 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e07bde2a-e592-4ac1-874b-3dc4a01d061d-serving-cert\") pod \"route-controller-manager-896d4d4c6-cr4gh\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.441819 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94b9d\" (UniqueName: \"kubernetes.io/projected/e07bde2a-e592-4ac1-874b-3dc4a01d061d-kube-api-access-94b9d\") pod \"route-controller-manager-896d4d4c6-cr4gh\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.441878 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-client-ca\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.441908 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-proxy-ca-bundles\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.441934 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a238584f-2a07-4a8c-b137-eea8509b7acf-serving-cert\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.441956 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-config\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.441987 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e07bde2a-e592-4ac1-874b-3dc4a01d061d-config\") pod \"route-controller-manager-896d4d4c6-cr4gh\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.442028 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e07bde2a-e592-4ac1-874b-3dc4a01d061d-client-ca\") pod \"route-controller-manager-896d4d4c6-cr4gh\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.442057 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c75z\" (UniqueName: \"kubernetes.io/projected/a238584f-2a07-4a8c-b137-eea8509b7acf-kube-api-access-8c75z\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.442897 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-client-ca\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.443531 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-proxy-ca-bundles\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.443693 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-config\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.444071 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e07bde2a-e592-4ac1-874b-3dc4a01d061d-client-ca\") pod \"route-controller-manager-896d4d4c6-cr4gh\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.445354 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e07bde2a-e592-4ac1-874b-3dc4a01d061d-config\") pod \"route-controller-manager-896d4d4c6-cr4gh\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.446136 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a238584f-2a07-4a8c-b137-eea8509b7acf-serving-cert\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.446162 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e07bde2a-e592-4ac1-874b-3dc4a01d061d-serving-cert\") pod \"route-controller-manager-896d4d4c6-cr4gh\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.464076 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c75z\" (UniqueName: \"kubernetes.io/projected/a238584f-2a07-4a8c-b137-eea8509b7acf-kube-api-access-8c75z\") pod \"controller-manager-6787ff745b-q4747\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.466519 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94b9d\" (UniqueName: \"kubernetes.io/projected/e07bde2a-e592-4ac1-874b-3dc4a01d061d-kube-api-access-94b9d\") pod \"route-controller-manager-896d4d4c6-cr4gh\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.535271 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.544900 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.922880 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6787ff745b-q4747"] Dec 11 08:42:41 crc kubenswrapper[4629]: I1211 08:42:41.980674 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh"] Dec 11 08:42:41 crc kubenswrapper[4629]: W1211 08:42:41.984753 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode07bde2a_e592_4ac1_874b_3dc4a01d061d.slice/crio-d657b6730a8ae83ac51df7bf5590341f8a9d0db26cbc093b1e505856fdfa822e WatchSource:0}: Error finding container d657b6730a8ae83ac51df7bf5590341f8a9d0db26cbc093b1e505856fdfa822e: Status 404 returned error can't find the container with id d657b6730a8ae83ac51df7bf5590341f8a9d0db26cbc093b1e505856fdfa822e Dec 11 08:42:42 crc kubenswrapper[4629]: I1211 08:42:42.208556 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbb99da1-1049-4e03-bc19-49988fc37947" path="/var/lib/kubelet/pods/dbb99da1-1049-4e03-bc19-49988fc37947/volumes" Dec 11 08:42:42 crc kubenswrapper[4629]: I1211 08:42:42.209517 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec0bf78c-4e35-4129-8976-c8d805991843" path="/var/lib/kubelet/pods/ec0bf78c-4e35-4129-8976-c8d805991843/volumes" Dec 11 08:42:42 crc kubenswrapper[4629]: I1211 08:42:42.941159 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" event={"ID":"a238584f-2a07-4a8c-b137-eea8509b7acf","Type":"ContainerStarted","Data":"d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46"} Dec 11 08:42:42 crc kubenswrapper[4629]: I1211 08:42:42.941209 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" event={"ID":"a238584f-2a07-4a8c-b137-eea8509b7acf","Type":"ContainerStarted","Data":"40ba8a3ec7dd4207aabbf0bea2eae426810740f47be0dc6d77d45170a9c18eb1"} Dec 11 08:42:42 crc kubenswrapper[4629]: I1211 08:42:42.941339 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:42 crc kubenswrapper[4629]: I1211 08:42:42.942490 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" event={"ID":"e07bde2a-e592-4ac1-874b-3dc4a01d061d","Type":"ContainerStarted","Data":"eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf"} Dec 11 08:42:42 crc kubenswrapper[4629]: I1211 08:42:42.942526 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" event={"ID":"e07bde2a-e592-4ac1-874b-3dc4a01d061d","Type":"ContainerStarted","Data":"d657b6730a8ae83ac51df7bf5590341f8a9d0db26cbc093b1e505856fdfa822e"} Dec 11 08:42:42 crc kubenswrapper[4629]: I1211 08:42:42.943025 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:42 crc kubenswrapper[4629]: I1211 08:42:42.947882 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:42 crc kubenswrapper[4629]: I1211 08:42:42.949331 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:42 crc kubenswrapper[4629]: I1211 08:42:42.956436 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" podStartSLOduration=1.9564091399999999 podStartE2EDuration="1.95640914s" podCreationTimestamp="2025-12-11 08:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:42:42.955553652 +0000 UTC m=+351.020969280" watchObservedRunningTime="2025-12-11 08:42:42.95640914 +0000 UTC m=+351.021824748" Dec 11 08:42:42 crc kubenswrapper[4629]: I1211 08:42:42.993466 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" podStartSLOduration=1.9934482340000002 podStartE2EDuration="1.993448234s" podCreationTimestamp="2025-12-11 08:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:42:42.992940467 +0000 UTC m=+351.058356075" watchObservedRunningTime="2025-12-11 08:42:42.993448234 +0000 UTC m=+351.058863842" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.486512 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.487226 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.561936 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-f8vg9"] Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.562933 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.579021 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-f8vg9"] Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.760331 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/38ae6fda-48dd-40f1-8678-47c934d69e0e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.760378 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/38ae6fda-48dd-40f1-8678-47c934d69e0e-registry-certificates\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.760403 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/38ae6fda-48dd-40f1-8678-47c934d69e0e-registry-tls\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.760522 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/38ae6fda-48dd-40f1-8678-47c934d69e0e-trusted-ca\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.760557 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwtf7\" (UniqueName: \"kubernetes.io/projected/38ae6fda-48dd-40f1-8678-47c934d69e0e-kube-api-access-hwtf7\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.760586 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/38ae6fda-48dd-40f1-8678-47c934d69e0e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.760605 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/38ae6fda-48dd-40f1-8678-47c934d69e0e-bound-sa-token\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.760657 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.796131 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.862346 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/38ae6fda-48dd-40f1-8678-47c934d69e0e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.862403 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/38ae6fda-48dd-40f1-8678-47c934d69e0e-registry-certificates\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.862433 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/38ae6fda-48dd-40f1-8678-47c934d69e0e-registry-tls\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.862499 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/38ae6fda-48dd-40f1-8678-47c934d69e0e-trusted-ca\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.862524 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwtf7\" (UniqueName: \"kubernetes.io/projected/38ae6fda-48dd-40f1-8678-47c934d69e0e-kube-api-access-hwtf7\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.862553 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/38ae6fda-48dd-40f1-8678-47c934d69e0e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.862579 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/38ae6fda-48dd-40f1-8678-47c934d69e0e-bound-sa-token\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.863084 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/38ae6fda-48dd-40f1-8678-47c934d69e0e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.863810 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/38ae6fda-48dd-40f1-8678-47c934d69e0e-trusted-ca\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.864551 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/38ae6fda-48dd-40f1-8678-47c934d69e0e-registry-certificates\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.876644 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/38ae6fda-48dd-40f1-8678-47c934d69e0e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.879195 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/38ae6fda-48dd-40f1-8678-47c934d69e0e-registry-tls\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.881132 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwtf7\" (UniqueName: \"kubernetes.io/projected/38ae6fda-48dd-40f1-8678-47c934d69e0e-kube-api-access-hwtf7\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.882636 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/38ae6fda-48dd-40f1-8678-47c934d69e0e-bound-sa-token\") pod \"image-registry-66df7c8f76-f8vg9\" (UID: \"38ae6fda-48dd-40f1-8678-47c934d69e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:50 crc kubenswrapper[4629]: I1211 08:42:50.892608 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:51 crc kubenswrapper[4629]: I1211 08:42:51.426648 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-f8vg9"] Dec 11 08:42:51 crc kubenswrapper[4629]: I1211 08:42:51.988134 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" event={"ID":"38ae6fda-48dd-40f1-8678-47c934d69e0e","Type":"ContainerStarted","Data":"70a29e0819cb6960a0f2d2c44d69aef18f424a7e183c63440696072d9e18915e"} Dec 11 08:42:51 crc kubenswrapper[4629]: I1211 08:42:51.988471 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" event={"ID":"38ae6fda-48dd-40f1-8678-47c934d69e0e","Type":"ContainerStarted","Data":"0b74dc44c382acef5dd30eb91e895b2e00c9d6edef754c120d4f812e565c2e32"} Dec 11 08:42:51 crc kubenswrapper[4629]: I1211 08:42:51.988495 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:42:52 crc kubenswrapper[4629]: I1211 08:42:52.006258 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" podStartSLOduration=2.006234557 podStartE2EDuration="2.006234557s" podCreationTimestamp="2025-12-11 08:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:42:52.003508816 +0000 UTC m=+360.068924434" watchObservedRunningTime="2025-12-11 08:42:52.006234557 +0000 UTC m=+360.071650155" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.092273 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6787ff745b-q4747"] Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.092919 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" podUID="a238584f-2a07-4a8c-b137-eea8509b7acf" containerName="controller-manager" containerID="cri-o://d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46" gracePeriod=30 Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.104977 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh"] Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.105509 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" podUID="e07bde2a-e592-4ac1-874b-3dc4a01d061d" containerName="route-controller-manager" containerID="cri-o://eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf" gracePeriod=30 Dec 11 08:42:56 crc kubenswrapper[4629]: E1211 08:42:56.261650 4629 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode07bde2a_e592_4ac1_874b_3dc4a01d061d.slice/crio-conmon-eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf.scope\": RecentStats: unable to find data in memory cache]" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.496513 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.632230 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94b9d\" (UniqueName: \"kubernetes.io/projected/e07bde2a-e592-4ac1-874b-3dc4a01d061d-kube-api-access-94b9d\") pod \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.633374 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e07bde2a-e592-4ac1-874b-3dc4a01d061d-config\") pod \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.633493 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e07bde2a-e592-4ac1-874b-3dc4a01d061d-client-ca\") pod \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.633529 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e07bde2a-e592-4ac1-874b-3dc4a01d061d-serving-cert\") pod \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\" (UID: \"e07bde2a-e592-4ac1-874b-3dc4a01d061d\") " Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.634142 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e07bde2a-e592-4ac1-874b-3dc4a01d061d-client-ca" (OuterVolumeSpecName: "client-ca") pod "e07bde2a-e592-4ac1-874b-3dc4a01d061d" (UID: "e07bde2a-e592-4ac1-874b-3dc4a01d061d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.634685 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e07bde2a-e592-4ac1-874b-3dc4a01d061d-config" (OuterVolumeSpecName: "config") pod "e07bde2a-e592-4ac1-874b-3dc4a01d061d" (UID: "e07bde2a-e592-4ac1-874b-3dc4a01d061d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.638114 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e07bde2a-e592-4ac1-874b-3dc4a01d061d-kube-api-access-94b9d" (OuterVolumeSpecName: "kube-api-access-94b9d") pod "e07bde2a-e592-4ac1-874b-3dc4a01d061d" (UID: "e07bde2a-e592-4ac1-874b-3dc4a01d061d"). InnerVolumeSpecName "kube-api-access-94b9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.644277 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e07bde2a-e592-4ac1-874b-3dc4a01d061d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e07bde2a-e592-4ac1-874b-3dc4a01d061d" (UID: "e07bde2a-e592-4ac1-874b-3dc4a01d061d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.705713 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.735518 4629 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e07bde2a-e592-4ac1-874b-3dc4a01d061d-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.735553 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e07bde2a-e592-4ac1-874b-3dc4a01d061d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.735563 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94b9d\" (UniqueName: \"kubernetes.io/projected/e07bde2a-e592-4ac1-874b-3dc4a01d061d-kube-api-access-94b9d\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.735573 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e07bde2a-e592-4ac1-874b-3dc4a01d061d-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.836697 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-proxy-ca-bundles\") pod \"a238584f-2a07-4a8c-b137-eea8509b7acf\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.836777 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a238584f-2a07-4a8c-b137-eea8509b7acf-serving-cert\") pod \"a238584f-2a07-4a8c-b137-eea8509b7acf\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.836804 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c75z\" (UniqueName: \"kubernetes.io/projected/a238584f-2a07-4a8c-b137-eea8509b7acf-kube-api-access-8c75z\") pod \"a238584f-2a07-4a8c-b137-eea8509b7acf\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.836827 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-client-ca\") pod \"a238584f-2a07-4a8c-b137-eea8509b7acf\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.836856 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-config\") pod \"a238584f-2a07-4a8c-b137-eea8509b7acf\" (UID: \"a238584f-2a07-4a8c-b137-eea8509b7acf\") " Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.837436 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a238584f-2a07-4a8c-b137-eea8509b7acf" (UID: "a238584f-2a07-4a8c-b137-eea8509b7acf"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.837540 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-client-ca" (OuterVolumeSpecName: "client-ca") pod "a238584f-2a07-4a8c-b137-eea8509b7acf" (UID: "a238584f-2a07-4a8c-b137-eea8509b7acf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.837692 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-config" (OuterVolumeSpecName: "config") pod "a238584f-2a07-4a8c-b137-eea8509b7acf" (UID: "a238584f-2a07-4a8c-b137-eea8509b7acf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.840516 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a238584f-2a07-4a8c-b137-eea8509b7acf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a238584f-2a07-4a8c-b137-eea8509b7acf" (UID: "a238584f-2a07-4a8c-b137-eea8509b7acf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.840631 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a238584f-2a07-4a8c-b137-eea8509b7acf-kube-api-access-8c75z" (OuterVolumeSpecName: "kube-api-access-8c75z") pod "a238584f-2a07-4a8c-b137-eea8509b7acf" (UID: "a238584f-2a07-4a8c-b137-eea8509b7acf"). InnerVolumeSpecName "kube-api-access-8c75z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.938603 4629 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.938671 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a238584f-2a07-4a8c-b137-eea8509b7acf-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.938684 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c75z\" (UniqueName: \"kubernetes.io/projected/a238584f-2a07-4a8c-b137-eea8509b7acf-kube-api-access-8c75z\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.938696 4629 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:56 crc kubenswrapper[4629]: I1211 08:42:56.938707 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a238584f-2a07-4a8c-b137-eea8509b7acf-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.012457 4629 generic.go:334] "Generic (PLEG): container finished" podID="a238584f-2a07-4a8c-b137-eea8509b7acf" containerID="d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46" exitCode=0 Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.012556 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.012548 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" event={"ID":"a238584f-2a07-4a8c-b137-eea8509b7acf","Type":"ContainerDied","Data":"d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46"} Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.012668 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6787ff745b-q4747" event={"ID":"a238584f-2a07-4a8c-b137-eea8509b7acf","Type":"ContainerDied","Data":"40ba8a3ec7dd4207aabbf0bea2eae426810740f47be0dc6d77d45170a9c18eb1"} Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.012695 4629 scope.go:117] "RemoveContainer" containerID="d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.013415 4629 generic.go:334] "Generic (PLEG): container finished" podID="e07bde2a-e592-4ac1-874b-3dc4a01d061d" containerID="eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf" exitCode=0 Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.013446 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" event={"ID":"e07bde2a-e592-4ac1-874b-3dc4a01d061d","Type":"ContainerDied","Data":"eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf"} Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.013472 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" event={"ID":"e07bde2a-e592-4ac1-874b-3dc4a01d061d","Type":"ContainerDied","Data":"d657b6730a8ae83ac51df7bf5590341f8a9d0db26cbc093b1e505856fdfa822e"} Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.013506 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.066285 4629 scope.go:117] "RemoveContainer" containerID="d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46" Dec 11 08:42:57 crc kubenswrapper[4629]: E1211 08:42:57.067411 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46\": container with ID starting with d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46 not found: ID does not exist" containerID="d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.067455 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46"} err="failed to get container status \"d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46\": rpc error: code = NotFound desc = could not find container \"d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46\": container with ID starting with d464259986f52717b2e5ae0482ec3fc55c470e607a963ea68014caad2c86ec46 not found: ID does not exist" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.067481 4629 scope.go:117] "RemoveContainer" containerID="eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.083336 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh"] Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.088579 4629 scope.go:117] "RemoveContainer" containerID="eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf" Dec 11 08:42:57 crc kubenswrapper[4629]: E1211 08:42:57.089143 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf\": container with ID starting with eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf not found: ID does not exist" containerID="eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.089366 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf"} err="failed to get container status \"eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf\": rpc error: code = NotFound desc = could not find container \"eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf\": container with ID starting with eaac76c20dca8e250d1eaf6f695b4bfbcb83b9b501c41b12b5e0dc8c24f5afbf not found: ID does not exist" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.089387 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-896d4d4c6-cr4gh"] Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.103078 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6787ff745b-q4747"] Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.105919 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6787ff745b-q4747"] Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.413077 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5c8985447-rmcrc"] Dec 11 08:42:57 crc kubenswrapper[4629]: E1211 08:42:57.413338 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a238584f-2a07-4a8c-b137-eea8509b7acf" containerName="controller-manager" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.413355 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a238584f-2a07-4a8c-b137-eea8509b7acf" containerName="controller-manager" Dec 11 08:42:57 crc kubenswrapper[4629]: E1211 08:42:57.413378 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e07bde2a-e592-4ac1-874b-3dc4a01d061d" containerName="route-controller-manager" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.413386 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="e07bde2a-e592-4ac1-874b-3dc4a01d061d" containerName="route-controller-manager" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.413503 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="e07bde2a-e592-4ac1-874b-3dc4a01d061d" containerName="route-controller-manager" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.413524 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="a238584f-2a07-4a8c-b137-eea8509b7acf" containerName="controller-manager" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.414003 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.418210 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.418773 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.418983 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.419157 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.419295 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.418979 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6"] Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.420172 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.420606 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.421930 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.421975 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.422221 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.422302 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.422339 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.424133 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.428506 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.431049 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c8985447-rmcrc"] Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.437578 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6"] Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.471743 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-client-ca\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.471802 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b6b8734-a403-49f1-8970-477e3c8e3955-serving-cert\") pod \"route-controller-manager-785fdfb9bc-46gx6\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.471898 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv27j\" (UniqueName: \"kubernetes.io/projected/5b6b8734-a403-49f1-8970-477e3c8e3955-kube-api-access-fv27j\") pod \"route-controller-manager-785fdfb9bc-46gx6\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.471990 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5b6b8734-a403-49f1-8970-477e3c8e3955-client-ca\") pod \"route-controller-manager-785fdfb9bc-46gx6\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.472016 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d17913a-3b0c-4347-97d7-9282f7ad24f0-serving-cert\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.472048 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-proxy-ca-bundles\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.472088 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b6b8734-a403-49f1-8970-477e3c8e3955-config\") pod \"route-controller-manager-785fdfb9bc-46gx6\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.472174 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjgdg\" (UniqueName: \"kubernetes.io/projected/9d17913a-3b0c-4347-97d7-9282f7ad24f0-kube-api-access-wjgdg\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.472206 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-config\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.572651 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjgdg\" (UniqueName: \"kubernetes.io/projected/9d17913a-3b0c-4347-97d7-9282f7ad24f0-kube-api-access-wjgdg\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.572702 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-config\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.572758 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-client-ca\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.572778 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b6b8734-a403-49f1-8970-477e3c8e3955-serving-cert\") pod \"route-controller-manager-785fdfb9bc-46gx6\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.572795 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv27j\" (UniqueName: \"kubernetes.io/projected/5b6b8734-a403-49f1-8970-477e3c8e3955-kube-api-access-fv27j\") pod \"route-controller-manager-785fdfb9bc-46gx6\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.572818 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5b6b8734-a403-49f1-8970-477e3c8e3955-client-ca\") pod \"route-controller-manager-785fdfb9bc-46gx6\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.572834 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d17913a-3b0c-4347-97d7-9282f7ad24f0-serving-cert\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.572873 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-proxy-ca-bundles\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.572900 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b6b8734-a403-49f1-8970-477e3c8e3955-config\") pod \"route-controller-manager-785fdfb9bc-46gx6\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.573881 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-client-ca\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.574210 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5b6b8734-a403-49f1-8970-477e3c8e3955-client-ca\") pod \"route-controller-manager-785fdfb9bc-46gx6\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.574364 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b6b8734-a403-49f1-8970-477e3c8e3955-config\") pod \"route-controller-manager-785fdfb9bc-46gx6\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.575147 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-proxy-ca-bundles\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.575182 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-config\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.577807 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b6b8734-a403-49f1-8970-477e3c8e3955-serving-cert\") pod \"route-controller-manager-785fdfb9bc-46gx6\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.578283 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d17913a-3b0c-4347-97d7-9282f7ad24f0-serving-cert\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.593271 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv27j\" (UniqueName: \"kubernetes.io/projected/5b6b8734-a403-49f1-8970-477e3c8e3955-kube-api-access-fv27j\") pod \"route-controller-manager-785fdfb9bc-46gx6\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.597762 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjgdg\" (UniqueName: \"kubernetes.io/projected/9d17913a-3b0c-4347-97d7-9282f7ad24f0-kube-api-access-wjgdg\") pod \"controller-manager-5c8985447-rmcrc\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.738731 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.750471 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.937018 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c8985447-rmcrc"] Dec 11 08:42:57 crc kubenswrapper[4629]: I1211 08:42:57.991236 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6"] Dec 11 08:42:57 crc kubenswrapper[4629]: W1211 08:42:57.996196 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b6b8734_a403_49f1_8970_477e3c8e3955.slice/crio-d49e0bd3c2be44b0157d9239c83182b980d72dc005bb5d9ec863f3fcc0fa09a7 WatchSource:0}: Error finding container d49e0bd3c2be44b0157d9239c83182b980d72dc005bb5d9ec863f3fcc0fa09a7: Status 404 returned error can't find the container with id d49e0bd3c2be44b0157d9239c83182b980d72dc005bb5d9ec863f3fcc0fa09a7 Dec 11 08:42:58 crc kubenswrapper[4629]: I1211 08:42:58.018564 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" event={"ID":"9d17913a-3b0c-4347-97d7-9282f7ad24f0","Type":"ContainerStarted","Data":"b1ff613520794878d266d0c7a9abd244c25eeb80477301fb63ba440e06025134"} Dec 11 08:42:58 crc kubenswrapper[4629]: I1211 08:42:58.019765 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" event={"ID":"5b6b8734-a403-49f1-8970-477e3c8e3955","Type":"ContainerStarted","Data":"d49e0bd3c2be44b0157d9239c83182b980d72dc005bb5d9ec863f3fcc0fa09a7"} Dec 11 08:42:58 crc kubenswrapper[4629]: I1211 08:42:58.215811 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a238584f-2a07-4a8c-b137-eea8509b7acf" path="/var/lib/kubelet/pods/a238584f-2a07-4a8c-b137-eea8509b7acf/volumes" Dec 11 08:42:58 crc kubenswrapper[4629]: I1211 08:42:58.218001 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e07bde2a-e592-4ac1-874b-3dc4a01d061d" path="/var/lib/kubelet/pods/e07bde2a-e592-4ac1-874b-3dc4a01d061d/volumes" Dec 11 08:42:59 crc kubenswrapper[4629]: I1211 08:42:59.028822 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" event={"ID":"9d17913a-3b0c-4347-97d7-9282f7ad24f0","Type":"ContainerStarted","Data":"30a7937a657a28e1244adbeaed21b599065643c325a20b90c88fb732623150f0"} Dec 11 08:42:59 crc kubenswrapper[4629]: I1211 08:42:59.029922 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:59 crc kubenswrapper[4629]: I1211 08:42:59.030420 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" event={"ID":"5b6b8734-a403-49f1-8970-477e3c8e3955","Type":"ContainerStarted","Data":"8728d2f61d4603efceb5fc1905b18ef27f0b23a1254d41dcba89349d2704e7ba"} Dec 11 08:42:59 crc kubenswrapper[4629]: I1211 08:42:59.031246 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:59 crc kubenswrapper[4629]: I1211 08:42:59.036177 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:42:59 crc kubenswrapper[4629]: I1211 08:42:59.037672 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:42:59 crc kubenswrapper[4629]: I1211 08:42:59.071872 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" podStartSLOduration=3.07183642 podStartE2EDuration="3.07183642s" podCreationTimestamp="2025-12-11 08:42:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:42:59.05144175 +0000 UTC m=+367.116857368" watchObservedRunningTime="2025-12-11 08:42:59.07183642 +0000 UTC m=+367.137252048" Dec 11 08:42:59 crc kubenswrapper[4629]: I1211 08:42:59.091473 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" podStartSLOduration=3.0914573340000002 podStartE2EDuration="3.091457334s" podCreationTimestamp="2025-12-11 08:42:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:42:59.088791305 +0000 UTC m=+367.154206913" watchObservedRunningTime="2025-12-11 08:42:59.091457334 +0000 UTC m=+367.156872942" Dec 11 08:43:10 crc kubenswrapper[4629]: I1211 08:43:10.898927 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-f8vg9" Dec 11 08:43:10 crc kubenswrapper[4629]: I1211 08:43:10.949269 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cmklq"] Dec 11 08:43:20 crc kubenswrapper[4629]: I1211 08:43:20.362833 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6"] Dec 11 08:43:20 crc kubenswrapper[4629]: I1211 08:43:20.363780 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" podUID="5b6b8734-a403-49f1-8970-477e3c8e3955" containerName="route-controller-manager" containerID="cri-o://8728d2f61d4603efceb5fc1905b18ef27f0b23a1254d41dcba89349d2704e7ba" gracePeriod=30 Dec 11 08:43:20 crc kubenswrapper[4629]: I1211 08:43:20.486190 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:43:20 crc kubenswrapper[4629]: I1211 08:43:20.486259 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.203908 4629 generic.go:334] "Generic (PLEG): container finished" podID="5b6b8734-a403-49f1-8970-477e3c8e3955" containerID="8728d2f61d4603efceb5fc1905b18ef27f0b23a1254d41dcba89349d2704e7ba" exitCode=0 Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.204223 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" event={"ID":"5b6b8734-a403-49f1-8970-477e3c8e3955","Type":"ContainerDied","Data":"8728d2f61d4603efceb5fc1905b18ef27f0b23a1254d41dcba89349d2704e7ba"} Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.518753 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.544885 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46"] Dec 11 08:43:21 crc kubenswrapper[4629]: E1211 08:43:21.545151 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b6b8734-a403-49f1-8970-477e3c8e3955" containerName="route-controller-manager" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.545166 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b6b8734-a403-49f1-8970-477e3c8e3955" containerName="route-controller-manager" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.545257 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b6b8734-a403-49f1-8970-477e3c8e3955" containerName="route-controller-manager" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.545684 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.564014 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46"] Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.666299 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b6b8734-a403-49f1-8970-477e3c8e3955-serving-cert\") pod \"5b6b8734-a403-49f1-8970-477e3c8e3955\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.666350 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b6b8734-a403-49f1-8970-477e3c8e3955-config\") pod \"5b6b8734-a403-49f1-8970-477e3c8e3955\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.666381 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5b6b8734-a403-49f1-8970-477e3c8e3955-client-ca\") pod \"5b6b8734-a403-49f1-8970-477e3c8e3955\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.666406 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv27j\" (UniqueName: \"kubernetes.io/projected/5b6b8734-a403-49f1-8970-477e3c8e3955-kube-api-access-fv27j\") pod \"5b6b8734-a403-49f1-8970-477e3c8e3955\" (UID: \"5b6b8734-a403-49f1-8970-477e3c8e3955\") " Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.666595 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l77p9\" (UniqueName: \"kubernetes.io/projected/8b26325c-ac22-4cf4-8a53-62068c910167-kube-api-access-l77p9\") pod \"route-controller-manager-896d4d4c6-6tl46\" (UID: \"8b26325c-ac22-4cf4-8a53-62068c910167\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.666644 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b26325c-ac22-4cf4-8a53-62068c910167-client-ca\") pod \"route-controller-manager-896d4d4c6-6tl46\" (UID: \"8b26325c-ac22-4cf4-8a53-62068c910167\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.666707 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b26325c-ac22-4cf4-8a53-62068c910167-serving-cert\") pod \"route-controller-manager-896d4d4c6-6tl46\" (UID: \"8b26325c-ac22-4cf4-8a53-62068c910167\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.666745 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b26325c-ac22-4cf4-8a53-62068c910167-config\") pod \"route-controller-manager-896d4d4c6-6tl46\" (UID: \"8b26325c-ac22-4cf4-8a53-62068c910167\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.667404 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b6b8734-a403-49f1-8970-477e3c8e3955-client-ca" (OuterVolumeSpecName: "client-ca") pod "5b6b8734-a403-49f1-8970-477e3c8e3955" (UID: "5b6b8734-a403-49f1-8970-477e3c8e3955"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.667418 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b6b8734-a403-49f1-8970-477e3c8e3955-config" (OuterVolumeSpecName: "config") pod "5b6b8734-a403-49f1-8970-477e3c8e3955" (UID: "5b6b8734-a403-49f1-8970-477e3c8e3955"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.673019 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b6b8734-a403-49f1-8970-477e3c8e3955-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5b6b8734-a403-49f1-8970-477e3c8e3955" (UID: "5b6b8734-a403-49f1-8970-477e3c8e3955"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.673743 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b6b8734-a403-49f1-8970-477e3c8e3955-kube-api-access-fv27j" (OuterVolumeSpecName: "kube-api-access-fv27j") pod "5b6b8734-a403-49f1-8970-477e3c8e3955" (UID: "5b6b8734-a403-49f1-8970-477e3c8e3955"). InnerVolumeSpecName "kube-api-access-fv27j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.768315 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b26325c-ac22-4cf4-8a53-62068c910167-serving-cert\") pod \"route-controller-manager-896d4d4c6-6tl46\" (UID: \"8b26325c-ac22-4cf4-8a53-62068c910167\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.768392 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b26325c-ac22-4cf4-8a53-62068c910167-config\") pod \"route-controller-manager-896d4d4c6-6tl46\" (UID: \"8b26325c-ac22-4cf4-8a53-62068c910167\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.768427 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l77p9\" (UniqueName: \"kubernetes.io/projected/8b26325c-ac22-4cf4-8a53-62068c910167-kube-api-access-l77p9\") pod \"route-controller-manager-896d4d4c6-6tl46\" (UID: \"8b26325c-ac22-4cf4-8a53-62068c910167\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.768462 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b26325c-ac22-4cf4-8a53-62068c910167-client-ca\") pod \"route-controller-manager-896d4d4c6-6tl46\" (UID: \"8b26325c-ac22-4cf4-8a53-62068c910167\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.768510 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b6b8734-a403-49f1-8970-477e3c8e3955-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.768521 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b6b8734-a403-49f1-8970-477e3c8e3955-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.768530 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv27j\" (UniqueName: \"kubernetes.io/projected/5b6b8734-a403-49f1-8970-477e3c8e3955-kube-api-access-fv27j\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.768540 4629 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5b6b8734-a403-49f1-8970-477e3c8e3955-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.769650 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b26325c-ac22-4cf4-8a53-62068c910167-client-ca\") pod \"route-controller-manager-896d4d4c6-6tl46\" (UID: \"8b26325c-ac22-4cf4-8a53-62068c910167\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.770495 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b26325c-ac22-4cf4-8a53-62068c910167-config\") pod \"route-controller-manager-896d4d4c6-6tl46\" (UID: \"8b26325c-ac22-4cf4-8a53-62068c910167\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.773727 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b26325c-ac22-4cf4-8a53-62068c910167-serving-cert\") pod \"route-controller-manager-896d4d4c6-6tl46\" (UID: \"8b26325c-ac22-4cf4-8a53-62068c910167\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.789404 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l77p9\" (UniqueName: \"kubernetes.io/projected/8b26325c-ac22-4cf4-8a53-62068c910167-kube-api-access-l77p9\") pod \"route-controller-manager-896d4d4c6-6tl46\" (UID: \"8b26325c-ac22-4cf4-8a53-62068c910167\") " pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:21 crc kubenswrapper[4629]: I1211 08:43:21.865676 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:22 crc kubenswrapper[4629]: I1211 08:43:22.212623 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" event={"ID":"5b6b8734-a403-49f1-8970-477e3c8e3955","Type":"ContainerDied","Data":"d49e0bd3c2be44b0157d9239c83182b980d72dc005bb5d9ec863f3fcc0fa09a7"} Dec 11 08:43:22 crc kubenswrapper[4629]: I1211 08:43:22.212684 4629 scope.go:117] "RemoveContainer" containerID="8728d2f61d4603efceb5fc1905b18ef27f0b23a1254d41dcba89349d2704e7ba" Dec 11 08:43:22 crc kubenswrapper[4629]: I1211 08:43:22.212820 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6" Dec 11 08:43:22 crc kubenswrapper[4629]: I1211 08:43:22.245102 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6"] Dec 11 08:43:22 crc kubenswrapper[4629]: I1211 08:43:22.248596 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-785fdfb9bc-46gx6"] Dec 11 08:43:22 crc kubenswrapper[4629]: I1211 08:43:22.272400 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46"] Dec 11 08:43:23 crc kubenswrapper[4629]: I1211 08:43:23.219940 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" event={"ID":"8b26325c-ac22-4cf4-8a53-62068c910167","Type":"ContainerStarted","Data":"2ab89539725b21bf36ddeaa0d8e975deae476686a6110033f860f3d6ceec3f33"} Dec 11 08:43:23 crc kubenswrapper[4629]: I1211 08:43:23.220352 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" event={"ID":"8b26325c-ac22-4cf4-8a53-62068c910167","Type":"ContainerStarted","Data":"eca125cdd30dde999002ed52cc0837a8af8f4028677bfbed7501e73ef141c9e4"} Dec 11 08:43:23 crc kubenswrapper[4629]: I1211 08:43:23.220373 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:23 crc kubenswrapper[4629]: I1211 08:43:23.225457 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" Dec 11 08:43:23 crc kubenswrapper[4629]: I1211 08:43:23.236935 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-896d4d4c6-6tl46" podStartSLOduration=3.236903591 podStartE2EDuration="3.236903591s" podCreationTimestamp="2025-12-11 08:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:43:23.234646406 +0000 UTC m=+391.300062024" watchObservedRunningTime="2025-12-11 08:43:23.236903591 +0000 UTC m=+391.302319209" Dec 11 08:43:24 crc kubenswrapper[4629]: I1211 08:43:24.204695 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b6b8734-a403-49f1-8970-477e3c8e3955" path="/var/lib/kubelet/pods/5b6b8734-a403-49f1-8970-477e3c8e3955/volumes" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.149567 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jrkpq"] Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.151173 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.154598 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.156959 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jrkpq"] Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.321616 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v7wv\" (UniqueName: \"kubernetes.io/projected/3f21f86f-d195-44ca-9403-d5407b466aff-kube-api-access-5v7wv\") pod \"redhat-operators-jrkpq\" (UID: \"3f21f86f-d195-44ca-9403-d5407b466aff\") " pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.321695 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f21f86f-d195-44ca-9403-d5407b466aff-utilities\") pod \"redhat-operators-jrkpq\" (UID: \"3f21f86f-d195-44ca-9403-d5407b466aff\") " pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.321769 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f21f86f-d195-44ca-9403-d5407b466aff-catalog-content\") pod \"redhat-operators-jrkpq\" (UID: \"3f21f86f-d195-44ca-9403-d5407b466aff\") " pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.347383 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k7x2k"] Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.348435 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.350924 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.363461 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7x2k"] Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.423246 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f21f86f-d195-44ca-9403-d5407b466aff-catalog-content\") pod \"redhat-operators-jrkpq\" (UID: \"3f21f86f-d195-44ca-9403-d5407b466aff\") " pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.423351 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5582d1-e74c-4c04-8370-202da8ffb11f-utilities\") pod \"redhat-marketplace-k7x2k\" (UID: \"2f5582d1-e74c-4c04-8370-202da8ffb11f\") " pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.423381 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km84g\" (UniqueName: \"kubernetes.io/projected/2f5582d1-e74c-4c04-8370-202da8ffb11f-kube-api-access-km84g\") pod \"redhat-marketplace-k7x2k\" (UID: \"2f5582d1-e74c-4c04-8370-202da8ffb11f\") " pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.423425 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v7wv\" (UniqueName: \"kubernetes.io/projected/3f21f86f-d195-44ca-9403-d5407b466aff-kube-api-access-5v7wv\") pod \"redhat-operators-jrkpq\" (UID: \"3f21f86f-d195-44ca-9403-d5407b466aff\") " pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.423493 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5582d1-e74c-4c04-8370-202da8ffb11f-catalog-content\") pod \"redhat-marketplace-k7x2k\" (UID: \"2f5582d1-e74c-4c04-8370-202da8ffb11f\") " pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.423530 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f21f86f-d195-44ca-9403-d5407b466aff-utilities\") pod \"redhat-operators-jrkpq\" (UID: \"3f21f86f-d195-44ca-9403-d5407b466aff\") " pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.423977 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f21f86f-d195-44ca-9403-d5407b466aff-catalog-content\") pod \"redhat-operators-jrkpq\" (UID: \"3f21f86f-d195-44ca-9403-d5407b466aff\") " pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.424046 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f21f86f-d195-44ca-9403-d5407b466aff-utilities\") pod \"redhat-operators-jrkpq\" (UID: \"3f21f86f-d195-44ca-9403-d5407b466aff\") " pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.441976 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v7wv\" (UniqueName: \"kubernetes.io/projected/3f21f86f-d195-44ca-9403-d5407b466aff-kube-api-access-5v7wv\") pod \"redhat-operators-jrkpq\" (UID: \"3f21f86f-d195-44ca-9403-d5407b466aff\") " pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.472560 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.525255 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5582d1-e74c-4c04-8370-202da8ffb11f-utilities\") pod \"redhat-marketplace-k7x2k\" (UID: \"2f5582d1-e74c-4c04-8370-202da8ffb11f\") " pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.525627 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km84g\" (UniqueName: \"kubernetes.io/projected/2f5582d1-e74c-4c04-8370-202da8ffb11f-kube-api-access-km84g\") pod \"redhat-marketplace-k7x2k\" (UID: \"2f5582d1-e74c-4c04-8370-202da8ffb11f\") " pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.525696 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5582d1-e74c-4c04-8370-202da8ffb11f-catalog-content\") pod \"redhat-marketplace-k7x2k\" (UID: \"2f5582d1-e74c-4c04-8370-202da8ffb11f\") " pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.525823 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5582d1-e74c-4c04-8370-202da8ffb11f-utilities\") pod \"redhat-marketplace-k7x2k\" (UID: \"2f5582d1-e74c-4c04-8370-202da8ffb11f\") " pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.526175 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5582d1-e74c-4c04-8370-202da8ffb11f-catalog-content\") pod \"redhat-marketplace-k7x2k\" (UID: \"2f5582d1-e74c-4c04-8370-202da8ffb11f\") " pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.553040 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km84g\" (UniqueName: \"kubernetes.io/projected/2f5582d1-e74c-4c04-8370-202da8ffb11f-kube-api-access-km84g\") pod \"redhat-marketplace-k7x2k\" (UID: \"2f5582d1-e74c-4c04-8370-202da8ffb11f\") " pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.664496 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:25 crc kubenswrapper[4629]: I1211 08:43:25.939333 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jrkpq"] Dec 11 08:43:26 crc kubenswrapper[4629]: I1211 08:43:26.043196 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7x2k"] Dec 11 08:43:26 crc kubenswrapper[4629]: W1211 08:43:26.047275 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f5582d1_e74c_4c04_8370_202da8ffb11f.slice/crio-de9de48d4450d9a56cf7655ee03a87dac72a7907378515ece044b3bd8f709915 WatchSource:0}: Error finding container de9de48d4450d9a56cf7655ee03a87dac72a7907378515ece044b3bd8f709915: Status 404 returned error can't find the container with id de9de48d4450d9a56cf7655ee03a87dac72a7907378515ece044b3bd8f709915 Dec 11 08:43:26 crc kubenswrapper[4629]: I1211 08:43:26.232878 4629 generic.go:334] "Generic (PLEG): container finished" podID="3f21f86f-d195-44ca-9403-d5407b466aff" containerID="64b10a14676ed874e600cc2cf955a0667def090de26f45d2d8771df7bd4173cd" exitCode=0 Dec 11 08:43:26 crc kubenswrapper[4629]: I1211 08:43:26.232961 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrkpq" event={"ID":"3f21f86f-d195-44ca-9403-d5407b466aff","Type":"ContainerDied","Data":"64b10a14676ed874e600cc2cf955a0667def090de26f45d2d8771df7bd4173cd"} Dec 11 08:43:26 crc kubenswrapper[4629]: I1211 08:43:26.233004 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrkpq" event={"ID":"3f21f86f-d195-44ca-9403-d5407b466aff","Type":"ContainerStarted","Data":"ad28313b04507a23bd03465e4c71d66e3df98a428c8642cb6c0abd223a7ef4b0"} Dec 11 08:43:26 crc kubenswrapper[4629]: I1211 08:43:26.234367 4629 generic.go:334] "Generic (PLEG): container finished" podID="2f5582d1-e74c-4c04-8370-202da8ffb11f" containerID="91123994f63d4162a3d07ae1e77ad26284016c1e061e68ccbd17a27edb7701a0" exitCode=0 Dec 11 08:43:26 crc kubenswrapper[4629]: I1211 08:43:26.234400 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7x2k" event={"ID":"2f5582d1-e74c-4c04-8370-202da8ffb11f","Type":"ContainerDied","Data":"91123994f63d4162a3d07ae1e77ad26284016c1e061e68ccbd17a27edb7701a0"} Dec 11 08:43:26 crc kubenswrapper[4629]: I1211 08:43:26.234421 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7x2k" event={"ID":"2f5582d1-e74c-4c04-8370-202da8ffb11f","Type":"ContainerStarted","Data":"de9de48d4450d9a56cf7655ee03a87dac72a7907378515ece044b3bd8f709915"} Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.241092 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrkpq" event={"ID":"3f21f86f-d195-44ca-9403-d5407b466aff","Type":"ContainerStarted","Data":"a1077db95987027ad702444c15060648858558b75c22a60072a68e8f51af3654"} Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.553672 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bv9ls"] Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.555237 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.557730 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.569089 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bv9ls"] Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.661964 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130-utilities\") pod \"certified-operators-bv9ls\" (UID: \"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130\") " pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.662585 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8bgc\" (UniqueName: \"kubernetes.io/projected/d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130-kube-api-access-r8bgc\") pod \"certified-operators-bv9ls\" (UID: \"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130\") " pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.662692 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130-catalog-content\") pod \"certified-operators-bv9ls\" (UID: \"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130\") " pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.746417 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lxnk8"] Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.747519 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.749598 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.774951 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130-utilities\") pod \"certified-operators-bv9ls\" (UID: \"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130\") " pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.775059 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8bgc\" (UniqueName: \"kubernetes.io/projected/d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130-kube-api-access-r8bgc\") pod \"certified-operators-bv9ls\" (UID: \"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130\") " pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.775110 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130-catalog-content\") pod \"certified-operators-bv9ls\" (UID: \"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130\") " pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.775688 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130-catalog-content\") pod \"certified-operators-bv9ls\" (UID: \"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130\") " pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.775728 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130-utilities\") pod \"certified-operators-bv9ls\" (UID: \"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130\") " pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.781915 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lxnk8"] Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.803405 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8bgc\" (UniqueName: \"kubernetes.io/projected/d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130-kube-api-access-r8bgc\") pod \"certified-operators-bv9ls\" (UID: \"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130\") " pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.870416 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.876306 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287e9dc5-e848-4c9c-b07e-14b244e606ef-catalog-content\") pod \"community-operators-lxnk8\" (UID: \"287e9dc5-e848-4c9c-b07e-14b244e606ef\") " pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.876460 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287e9dc5-e848-4c9c-b07e-14b244e606ef-utilities\") pod \"community-operators-lxnk8\" (UID: \"287e9dc5-e848-4c9c-b07e-14b244e606ef\") " pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.876515 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqfkf\" (UniqueName: \"kubernetes.io/projected/287e9dc5-e848-4c9c-b07e-14b244e606ef-kube-api-access-kqfkf\") pod \"community-operators-lxnk8\" (UID: \"287e9dc5-e848-4c9c-b07e-14b244e606ef\") " pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.978298 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287e9dc5-e848-4c9c-b07e-14b244e606ef-catalog-content\") pod \"community-operators-lxnk8\" (UID: \"287e9dc5-e848-4c9c-b07e-14b244e606ef\") " pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.978659 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287e9dc5-e848-4c9c-b07e-14b244e606ef-utilities\") pod \"community-operators-lxnk8\" (UID: \"287e9dc5-e848-4c9c-b07e-14b244e606ef\") " pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.978718 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqfkf\" (UniqueName: \"kubernetes.io/projected/287e9dc5-e848-4c9c-b07e-14b244e606ef-kube-api-access-kqfkf\") pod \"community-operators-lxnk8\" (UID: \"287e9dc5-e848-4c9c-b07e-14b244e606ef\") " pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.978890 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287e9dc5-e848-4c9c-b07e-14b244e606ef-catalog-content\") pod \"community-operators-lxnk8\" (UID: \"287e9dc5-e848-4c9c-b07e-14b244e606ef\") " pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.979162 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287e9dc5-e848-4c9c-b07e-14b244e606ef-utilities\") pod \"community-operators-lxnk8\" (UID: \"287e9dc5-e848-4c9c-b07e-14b244e606ef\") " pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:27 crc kubenswrapper[4629]: I1211 08:43:27.996885 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqfkf\" (UniqueName: \"kubernetes.io/projected/287e9dc5-e848-4c9c-b07e-14b244e606ef-kube-api-access-kqfkf\") pod \"community-operators-lxnk8\" (UID: \"287e9dc5-e848-4c9c-b07e-14b244e606ef\") " pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:28 crc kubenswrapper[4629]: I1211 08:43:28.063608 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:28 crc kubenswrapper[4629]: I1211 08:43:28.251188 4629 generic.go:334] "Generic (PLEG): container finished" podID="2f5582d1-e74c-4c04-8370-202da8ffb11f" containerID="68b6c36f59934a25b2686fd4076650bac295ec892378f6fe841c338534e4e43f" exitCode=0 Dec 11 08:43:28 crc kubenswrapper[4629]: I1211 08:43:28.251261 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7x2k" event={"ID":"2f5582d1-e74c-4c04-8370-202da8ffb11f","Type":"ContainerDied","Data":"68b6c36f59934a25b2686fd4076650bac295ec892378f6fe841c338534e4e43f"} Dec 11 08:43:28 crc kubenswrapper[4629]: I1211 08:43:28.261255 4629 generic.go:334] "Generic (PLEG): container finished" podID="3f21f86f-d195-44ca-9403-d5407b466aff" containerID="a1077db95987027ad702444c15060648858558b75c22a60072a68e8f51af3654" exitCode=0 Dec 11 08:43:28 crc kubenswrapper[4629]: I1211 08:43:28.261298 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrkpq" event={"ID":"3f21f86f-d195-44ca-9403-d5407b466aff","Type":"ContainerDied","Data":"a1077db95987027ad702444c15060648858558b75c22a60072a68e8f51af3654"} Dec 11 08:43:28 crc kubenswrapper[4629]: I1211 08:43:28.291073 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bv9ls"] Dec 11 08:43:28 crc kubenswrapper[4629]: W1211 08:43:28.297161 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1bcb2f9_52e1_4c00_a674_fb1b1b9e8130.slice/crio-c68fd90ad329adf69c37f3c215abeeb46a4091460b5d8a8a63598a5ec66e8503 WatchSource:0}: Error finding container c68fd90ad329adf69c37f3c215abeeb46a4091460b5d8a8a63598a5ec66e8503: Status 404 returned error can't find the container with id c68fd90ad329adf69c37f3c215abeeb46a4091460b5d8a8a63598a5ec66e8503 Dec 11 08:43:28 crc kubenswrapper[4629]: I1211 08:43:28.526200 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lxnk8"] Dec 11 08:43:28 crc kubenswrapper[4629]: W1211 08:43:28.527270 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod287e9dc5_e848_4c9c_b07e_14b244e606ef.slice/crio-5bafc84b43b141fce7ff93c0d4fda3cc724f1ec0153fe6341a83a0f433439fbf WatchSource:0}: Error finding container 5bafc84b43b141fce7ff93c0d4fda3cc724f1ec0153fe6341a83a0f433439fbf: Status 404 returned error can't find the container with id 5bafc84b43b141fce7ff93c0d4fda3cc724f1ec0153fe6341a83a0f433439fbf Dec 11 08:43:29 crc kubenswrapper[4629]: I1211 08:43:29.280526 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7x2k" event={"ID":"2f5582d1-e74c-4c04-8370-202da8ffb11f","Type":"ContainerStarted","Data":"9eae018293731500c1814eba952cf06f647ad119288774ecaadab90467ce6039"} Dec 11 08:43:29 crc kubenswrapper[4629]: I1211 08:43:29.285731 4629 generic.go:334] "Generic (PLEG): container finished" podID="287e9dc5-e848-4c9c-b07e-14b244e606ef" containerID="d0ec57f3bbb70a2a7d73829d700e5515f536f0981400565142ac8a2d6cfbf786" exitCode=0 Dec 11 08:43:29 crc kubenswrapper[4629]: I1211 08:43:29.285778 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxnk8" event={"ID":"287e9dc5-e848-4c9c-b07e-14b244e606ef","Type":"ContainerDied","Data":"d0ec57f3bbb70a2a7d73829d700e5515f536f0981400565142ac8a2d6cfbf786"} Dec 11 08:43:29 crc kubenswrapper[4629]: I1211 08:43:29.285811 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxnk8" event={"ID":"287e9dc5-e848-4c9c-b07e-14b244e606ef","Type":"ContainerStarted","Data":"5bafc84b43b141fce7ff93c0d4fda3cc724f1ec0153fe6341a83a0f433439fbf"} Dec 11 08:43:29 crc kubenswrapper[4629]: I1211 08:43:29.287788 4629 generic.go:334] "Generic (PLEG): container finished" podID="d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130" containerID="62f83090ff024cf3663a9b98b65df4719d1a6ae07d322dfbdf8eeb3cfff9c473" exitCode=0 Dec 11 08:43:29 crc kubenswrapper[4629]: I1211 08:43:29.287896 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bv9ls" event={"ID":"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130","Type":"ContainerDied","Data":"62f83090ff024cf3663a9b98b65df4719d1a6ae07d322dfbdf8eeb3cfff9c473"} Dec 11 08:43:29 crc kubenswrapper[4629]: I1211 08:43:29.287953 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bv9ls" event={"ID":"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130","Type":"ContainerStarted","Data":"c68fd90ad329adf69c37f3c215abeeb46a4091460b5d8a8a63598a5ec66e8503"} Dec 11 08:43:29 crc kubenswrapper[4629]: I1211 08:43:29.291282 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrkpq" event={"ID":"3f21f86f-d195-44ca-9403-d5407b466aff","Type":"ContainerStarted","Data":"d18807c90d87e44c4f1902070079b1ca8afd15effa192b8eeadd336beb1b80f3"} Dec 11 08:43:29 crc kubenswrapper[4629]: I1211 08:43:29.315200 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k7x2k" podStartSLOduration=1.796726118 podStartE2EDuration="4.315178821s" podCreationTimestamp="2025-12-11 08:43:25 +0000 UTC" firstStartedPulling="2025-12-11 08:43:26.235537434 +0000 UTC m=+394.300953032" lastFinishedPulling="2025-12-11 08:43:28.753990127 +0000 UTC m=+396.819405735" observedRunningTime="2025-12-11 08:43:29.312602074 +0000 UTC m=+397.378017682" watchObservedRunningTime="2025-12-11 08:43:29.315178821 +0000 UTC m=+397.380594429" Dec 11 08:43:29 crc kubenswrapper[4629]: I1211 08:43:29.378054 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jrkpq" podStartSLOduration=1.88016494 podStartE2EDuration="4.378037786s" podCreationTimestamp="2025-12-11 08:43:25 +0000 UTC" firstStartedPulling="2025-12-11 08:43:26.233921906 +0000 UTC m=+394.299337504" lastFinishedPulling="2025-12-11 08:43:28.731794742 +0000 UTC m=+396.797210350" observedRunningTime="2025-12-11 08:43:29.374741177 +0000 UTC m=+397.440156785" watchObservedRunningTime="2025-12-11 08:43:29.378037786 +0000 UTC m=+397.443453394" Dec 11 08:43:31 crc kubenswrapper[4629]: I1211 08:43:31.303407 4629 generic.go:334] "Generic (PLEG): container finished" podID="d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130" containerID="4c1aa351862f760a1d03f45bb2fa7a059c4f586dbe5b7b4a5b5c69d65eab199b" exitCode=0 Dec 11 08:43:31 crc kubenswrapper[4629]: I1211 08:43:31.303514 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bv9ls" event={"ID":"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130","Type":"ContainerDied","Data":"4c1aa351862f760a1d03f45bb2fa7a059c4f586dbe5b7b4a5b5c69d65eab199b"} Dec 11 08:43:31 crc kubenswrapper[4629]: I1211 08:43:31.306182 4629 generic.go:334] "Generic (PLEG): container finished" podID="287e9dc5-e848-4c9c-b07e-14b244e606ef" containerID="7f6e382579b3126e005ad23ce1bdcc87d501a1f6bf02b95df10217b2d7b8fb12" exitCode=0 Dec 11 08:43:31 crc kubenswrapper[4629]: I1211 08:43:31.306226 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxnk8" event={"ID":"287e9dc5-e848-4c9c-b07e-14b244e606ef","Type":"ContainerDied","Data":"7f6e382579b3126e005ad23ce1bdcc87d501a1f6bf02b95df10217b2d7b8fb12"} Dec 11 08:43:32 crc kubenswrapper[4629]: I1211 08:43:32.312798 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxnk8" event={"ID":"287e9dc5-e848-4c9c-b07e-14b244e606ef","Type":"ContainerStarted","Data":"50c6c961192a73c732b775764b347c9edc43600c06242915fcf2e05fdd493ce9"} Dec 11 08:43:32 crc kubenswrapper[4629]: I1211 08:43:32.315648 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bv9ls" event={"ID":"d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130","Type":"ContainerStarted","Data":"ce91852b587bf470762b0fd55fc1eda833cb9cb8ff214ef70091bcdea9e5c9f8"} Dec 11 08:43:32 crc kubenswrapper[4629]: I1211 08:43:32.333459 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lxnk8" podStartSLOduration=2.861862821 podStartE2EDuration="5.333438038s" podCreationTimestamp="2025-12-11 08:43:27 +0000 UTC" firstStartedPulling="2025-12-11 08:43:29.287316026 +0000 UTC m=+397.352731674" lastFinishedPulling="2025-12-11 08:43:31.758891283 +0000 UTC m=+399.824306891" observedRunningTime="2025-12-11 08:43:32.327955754 +0000 UTC m=+400.393371372" watchObservedRunningTime="2025-12-11 08:43:32.333438038 +0000 UTC m=+400.398853646" Dec 11 08:43:32 crc kubenswrapper[4629]: I1211 08:43:32.348994 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bv9ls" podStartSLOduration=2.8577296260000002 podStartE2EDuration="5.348974483s" podCreationTimestamp="2025-12-11 08:43:27 +0000 UTC" firstStartedPulling="2025-12-11 08:43:29.288966526 +0000 UTC m=+397.354382134" lastFinishedPulling="2025-12-11 08:43:31.780211383 +0000 UTC m=+399.845626991" observedRunningTime="2025-12-11 08:43:32.345781248 +0000 UTC m=+400.411196866" watchObservedRunningTime="2025-12-11 08:43:32.348974483 +0000 UTC m=+400.414390091" Dec 11 08:43:35 crc kubenswrapper[4629]: I1211 08:43:35.472730 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:35 crc kubenswrapper[4629]: I1211 08:43:35.473162 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:35 crc kubenswrapper[4629]: I1211 08:43:35.518699 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:35 crc kubenswrapper[4629]: I1211 08:43:35.665641 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:35 crc kubenswrapper[4629]: I1211 08:43:35.665686 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:35 crc kubenswrapper[4629]: I1211 08:43:35.706915 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:36 crc kubenswrapper[4629]: I1211 08:43:36.007351 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" podUID="caa0ac44-0445-4d00-af37-e00e3c3d7d5d" containerName="registry" containerID="cri-o://9e68593d9990ab48b82fe2cb1e46fcd4f08708f413c80eef3d21cb3fe55e5342" gracePeriod=30 Dec 11 08:43:36 crc kubenswrapper[4629]: I1211 08:43:36.391442 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k7x2k" Dec 11 08:43:36 crc kubenswrapper[4629]: I1211 08:43:36.392285 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jrkpq" Dec 11 08:43:37 crc kubenswrapper[4629]: I1211 08:43:37.871209 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:37 crc kubenswrapper[4629]: I1211 08:43:37.871286 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:37 crc kubenswrapper[4629]: I1211 08:43:37.908266 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:38 crc kubenswrapper[4629]: I1211 08:43:38.064543 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:38 crc kubenswrapper[4629]: I1211 08:43:38.064645 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:38 crc kubenswrapper[4629]: I1211 08:43:38.100517 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:38 crc kubenswrapper[4629]: I1211 08:43:38.366163 4629 generic.go:334] "Generic (PLEG): container finished" podID="caa0ac44-0445-4d00-af37-e00e3c3d7d5d" containerID="9e68593d9990ab48b82fe2cb1e46fcd4f08708f413c80eef3d21cb3fe55e5342" exitCode=0 Dec 11 08:43:38 crc kubenswrapper[4629]: I1211 08:43:38.366242 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" event={"ID":"caa0ac44-0445-4d00-af37-e00e3c3d7d5d","Type":"ContainerDied","Data":"9e68593d9990ab48b82fe2cb1e46fcd4f08708f413c80eef3d21cb3fe55e5342"} Dec 11 08:43:38 crc kubenswrapper[4629]: I1211 08:43:38.400744 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bv9ls" Dec 11 08:43:38 crc kubenswrapper[4629]: I1211 08:43:38.405281 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lxnk8" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.210637 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.372494 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.372960 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cmklq" event={"ID":"caa0ac44-0445-4d00-af37-e00e3c3d7d5d","Type":"ContainerDied","Data":"6c86c600db1b1ec60c55f57da586a335af3a81c694d79c75a575e5accc60ee30"} Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.372999 4629 scope.go:117] "RemoveContainer" containerID="9e68593d9990ab48b82fe2cb1e46fcd4f08708f413c80eef3d21cb3fe55e5342" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.376136 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-registry-certificates\") pod \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.376165 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vb6j5\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-kube-api-access-vb6j5\") pod \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.376214 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-ca-trust-extracted\") pod \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.376252 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-trusted-ca\") pod \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.376620 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.376646 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-installation-pull-secrets\") pod \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.376669 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-registry-tls\") pod \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.376684 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-bound-sa-token\") pod \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\" (UID: \"caa0ac44-0445-4d00-af37-e00e3c3d7d5d\") " Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.378114 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "caa0ac44-0445-4d00-af37-e00e3c3d7d5d" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.378953 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "caa0ac44-0445-4d00-af37-e00e3c3d7d5d" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.387079 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "caa0ac44-0445-4d00-af37-e00e3c3d7d5d" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.387365 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "caa0ac44-0445-4d00-af37-e00e3c3d7d5d" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.388340 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "caa0ac44-0445-4d00-af37-e00e3c3d7d5d" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.388531 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "caa0ac44-0445-4d00-af37-e00e3c3d7d5d" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.388710 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-kube-api-access-vb6j5" (OuterVolumeSpecName: "kube-api-access-vb6j5") pod "caa0ac44-0445-4d00-af37-e00e3c3d7d5d" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d"). InnerVolumeSpecName "kube-api-access-vb6j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.397147 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "caa0ac44-0445-4d00-af37-e00e3c3d7d5d" (UID: "caa0ac44-0445-4d00-af37-e00e3c3d7d5d"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.478659 4629 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.478962 4629 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.478975 4629 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.478988 4629 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.478997 4629 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.479006 4629 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.479014 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vb6j5\" (UniqueName: \"kubernetes.io/projected/caa0ac44-0445-4d00-af37-e00e3c3d7d5d-kube-api-access-vb6j5\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.706585 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cmklq"] Dec 11 08:43:39 crc kubenswrapper[4629]: I1211 08:43:39.713364 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cmklq"] Dec 11 08:43:40 crc kubenswrapper[4629]: I1211 08:43:40.192626 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5c8985447-rmcrc"] Dec 11 08:43:40 crc kubenswrapper[4629]: I1211 08:43:40.193262 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" podUID="9d17913a-3b0c-4347-97d7-9282f7ad24f0" containerName="controller-manager" containerID="cri-o://30a7937a657a28e1244adbeaed21b599065643c325a20b90c88fb732623150f0" gracePeriod=30 Dec 11 08:43:40 crc kubenswrapper[4629]: I1211 08:43:40.208611 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caa0ac44-0445-4d00-af37-e00e3c3d7d5d" path="/var/lib/kubelet/pods/caa0ac44-0445-4d00-af37-e00e3c3d7d5d/volumes" Dec 11 08:43:41 crc kubenswrapper[4629]: I1211 08:43:41.383696 4629 generic.go:334] "Generic (PLEG): container finished" podID="9d17913a-3b0c-4347-97d7-9282f7ad24f0" containerID="30a7937a657a28e1244adbeaed21b599065643c325a20b90c88fb732623150f0" exitCode=0 Dec 11 08:43:41 crc kubenswrapper[4629]: I1211 08:43:41.383747 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" event={"ID":"9d17913a-3b0c-4347-97d7-9282f7ad24f0","Type":"ContainerDied","Data":"30a7937a657a28e1244adbeaed21b599065643c325a20b90c88fb732623150f0"} Dec 11 08:43:41 crc kubenswrapper[4629]: I1211 08:43:41.822607 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:43:41 crc kubenswrapper[4629]: I1211 08:43:41.852427 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6787ff745b-p4jnz"] Dec 11 08:43:41 crc kubenswrapper[4629]: E1211 08:43:41.852673 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d17913a-3b0c-4347-97d7-9282f7ad24f0" containerName="controller-manager" Dec 11 08:43:41 crc kubenswrapper[4629]: I1211 08:43:41.852685 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d17913a-3b0c-4347-97d7-9282f7ad24f0" containerName="controller-manager" Dec 11 08:43:41 crc kubenswrapper[4629]: E1211 08:43:41.852696 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa0ac44-0445-4d00-af37-e00e3c3d7d5d" containerName="registry" Dec 11 08:43:41 crc kubenswrapper[4629]: I1211 08:43:41.852702 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa0ac44-0445-4d00-af37-e00e3c3d7d5d" containerName="registry" Dec 11 08:43:41 crc kubenswrapper[4629]: I1211 08:43:41.852793 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="caa0ac44-0445-4d00-af37-e00e3c3d7d5d" containerName="registry" Dec 11 08:43:41 crc kubenswrapper[4629]: I1211 08:43:41.852808 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d17913a-3b0c-4347-97d7-9282f7ad24f0" containerName="controller-manager" Dec 11 08:43:41 crc kubenswrapper[4629]: I1211 08:43:41.853244 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:41 crc kubenswrapper[4629]: I1211 08:43:41.873259 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6787ff745b-p4jnz"] Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.006798 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-proxy-ca-bundles\") pod \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.006891 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-config\") pod \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.006964 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d17913a-3b0c-4347-97d7-9282f7ad24f0-serving-cert\") pod \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.007009 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjgdg\" (UniqueName: \"kubernetes.io/projected/9d17913a-3b0c-4347-97d7-9282f7ad24f0-kube-api-access-wjgdg\") pod \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.007050 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-client-ca\") pod \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\" (UID: \"9d17913a-3b0c-4347-97d7-9282f7ad24f0\") " Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.007238 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqj5l\" (UniqueName: \"kubernetes.io/projected/0e95ae72-e3ad-48a2-91fa-a44469d955bd-kube-api-access-nqj5l\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.007269 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e95ae72-e3ad-48a2-91fa-a44469d955bd-serving-cert\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.007308 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0e95ae72-e3ad-48a2-91fa-a44469d955bd-proxy-ca-bundles\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.007337 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e95ae72-e3ad-48a2-91fa-a44469d955bd-config\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.007361 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0e95ae72-e3ad-48a2-91fa-a44469d955bd-client-ca\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.007755 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9d17913a-3b0c-4347-97d7-9282f7ad24f0" (UID: "9d17913a-3b0c-4347-97d7-9282f7ad24f0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.007926 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-config" (OuterVolumeSpecName: "config") pod "9d17913a-3b0c-4347-97d7-9282f7ad24f0" (UID: "9d17913a-3b0c-4347-97d7-9282f7ad24f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.008379 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-client-ca" (OuterVolumeSpecName: "client-ca") pod "9d17913a-3b0c-4347-97d7-9282f7ad24f0" (UID: "9d17913a-3b0c-4347-97d7-9282f7ad24f0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.013212 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d17913a-3b0c-4347-97d7-9282f7ad24f0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d17913a-3b0c-4347-97d7-9282f7ad24f0" (UID: "9d17913a-3b0c-4347-97d7-9282f7ad24f0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.013270 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d17913a-3b0c-4347-97d7-9282f7ad24f0-kube-api-access-wjgdg" (OuterVolumeSpecName: "kube-api-access-wjgdg") pod "9d17913a-3b0c-4347-97d7-9282f7ad24f0" (UID: "9d17913a-3b0c-4347-97d7-9282f7ad24f0"). InnerVolumeSpecName "kube-api-access-wjgdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.108517 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqj5l\" (UniqueName: \"kubernetes.io/projected/0e95ae72-e3ad-48a2-91fa-a44469d955bd-kube-api-access-nqj5l\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.108589 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e95ae72-e3ad-48a2-91fa-a44469d955bd-serving-cert\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.108624 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0e95ae72-e3ad-48a2-91fa-a44469d955bd-proxy-ca-bundles\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.108654 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e95ae72-e3ad-48a2-91fa-a44469d955bd-config\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.108673 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0e95ae72-e3ad-48a2-91fa-a44469d955bd-client-ca\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.108718 4629 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.108729 4629 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.108738 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d17913a-3b0c-4347-97d7-9282f7ad24f0-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.108748 4629 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d17913a-3b0c-4347-97d7-9282f7ad24f0-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.108757 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjgdg\" (UniqueName: \"kubernetes.io/projected/9d17913a-3b0c-4347-97d7-9282f7ad24f0-kube-api-access-wjgdg\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.109929 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0e95ae72-e3ad-48a2-91fa-a44469d955bd-client-ca\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.110256 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0e95ae72-e3ad-48a2-91fa-a44469d955bd-proxy-ca-bundles\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.111137 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e95ae72-e3ad-48a2-91fa-a44469d955bd-config\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.116375 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e95ae72-e3ad-48a2-91fa-a44469d955bd-serving-cert\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.124736 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqj5l\" (UniqueName: \"kubernetes.io/projected/0e95ae72-e3ad-48a2-91fa-a44469d955bd-kube-api-access-nqj5l\") pod \"controller-manager-6787ff745b-p4jnz\" (UID: \"0e95ae72-e3ad-48a2-91fa-a44469d955bd\") " pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.182195 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.393816 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" event={"ID":"9d17913a-3b0c-4347-97d7-9282f7ad24f0","Type":"ContainerDied","Data":"b1ff613520794878d266d0c7a9abd244c25eeb80477301fb63ba440e06025134"} Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.393908 4629 scope.go:117] "RemoveContainer" containerID="30a7937a657a28e1244adbeaed21b599065643c325a20b90c88fb732623150f0" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.393913 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c8985447-rmcrc" Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.417468 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5c8985447-rmcrc"] Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.420121 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5c8985447-rmcrc"] Dec 11 08:43:42 crc kubenswrapper[4629]: I1211 08:43:42.561330 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6787ff745b-p4jnz"] Dec 11 08:43:42 crc kubenswrapper[4629]: W1211 08:43:42.565265 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e95ae72_e3ad_48a2_91fa_a44469d955bd.slice/crio-b9b84ff75b0a30488363258d07f5bfe4f73a9b9dbb377211d6a49d3bc06ed10f WatchSource:0}: Error finding container b9b84ff75b0a30488363258d07f5bfe4f73a9b9dbb377211d6a49d3bc06ed10f: Status 404 returned error can't find the container with id b9b84ff75b0a30488363258d07f5bfe4f73a9b9dbb377211d6a49d3bc06ed10f Dec 11 08:43:43 crc kubenswrapper[4629]: I1211 08:43:43.401910 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" event={"ID":"0e95ae72-e3ad-48a2-91fa-a44469d955bd","Type":"ContainerStarted","Data":"afeb5d454311fec32359837294c66da29fe92672982b25aa4789d6bfd4d1360d"} Dec 11 08:43:43 crc kubenswrapper[4629]: I1211 08:43:43.401956 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" event={"ID":"0e95ae72-e3ad-48a2-91fa-a44469d955bd","Type":"ContainerStarted","Data":"b9b84ff75b0a30488363258d07f5bfe4f73a9b9dbb377211d6a49d3bc06ed10f"} Dec 11 08:43:43 crc kubenswrapper[4629]: I1211 08:43:43.402392 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:43 crc kubenswrapper[4629]: I1211 08:43:43.406413 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" Dec 11 08:43:43 crc kubenswrapper[4629]: I1211 08:43:43.421760 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6787ff745b-p4jnz" podStartSLOduration=3.421734022 podStartE2EDuration="3.421734022s" podCreationTimestamp="2025-12-11 08:43:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:43:43.419532407 +0000 UTC m=+411.484948015" watchObservedRunningTime="2025-12-11 08:43:43.421734022 +0000 UTC m=+411.487149630" Dec 11 08:43:44 crc kubenswrapper[4629]: I1211 08:43:44.206628 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d17913a-3b0c-4347-97d7-9282f7ad24f0" path="/var/lib/kubelet/pods/9d17913a-3b0c-4347-97d7-9282f7ad24f0/volumes" Dec 11 08:43:50 crc kubenswrapper[4629]: I1211 08:43:50.486648 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:43:50 crc kubenswrapper[4629]: I1211 08:43:50.487094 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:43:50 crc kubenswrapper[4629]: I1211 08:43:50.487154 4629 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:43:50 crc kubenswrapper[4629]: I1211 08:43:50.487784 4629 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ba210253304f5487d01920ea72fb77a38de3a8bfc04263290603cecc0511379"} pod="openshift-machine-config-operator/machine-config-daemon-wbld9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:43:50 crc kubenswrapper[4629]: I1211 08:43:50.487853 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" containerID="cri-o://8ba210253304f5487d01920ea72fb77a38de3a8bfc04263290603cecc0511379" gracePeriod=600 Dec 11 08:43:51 crc kubenswrapper[4629]: I1211 08:43:51.441421 4629 generic.go:334] "Generic (PLEG): container finished" podID="847b50b9-0093-404f-b0dd-63c814d8986b" containerID="8ba210253304f5487d01920ea72fb77a38de3a8bfc04263290603cecc0511379" exitCode=0 Dec 11 08:43:51 crc kubenswrapper[4629]: I1211 08:43:51.441494 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerDied","Data":"8ba210253304f5487d01920ea72fb77a38de3a8bfc04263290603cecc0511379"} Dec 11 08:43:51 crc kubenswrapper[4629]: I1211 08:43:51.441912 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"256bede4f7d04c63baf8ee22be7f6412368a168b18567d6d4b73acc6556f3040"} Dec 11 08:43:51 crc kubenswrapper[4629]: I1211 08:43:51.441931 4629 scope.go:117] "RemoveContainer" containerID="df0d293a2bb5ff9e4e52efc917093fefb328b6608a1293cdcbd49b88019a64a7" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.176387 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q"] Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.177946 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.181121 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.182118 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.185724 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-config-volume\") pod \"collect-profiles-29424045-lkk6q\" (UID: \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.185884 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-secret-volume\") pod \"collect-profiles-29424045-lkk6q\" (UID: \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.185938 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q9xm\" (UniqueName: \"kubernetes.io/projected/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-kube-api-access-7q9xm\") pod \"collect-profiles-29424045-lkk6q\" (UID: \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.191671 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q"] Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.287149 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-config-volume\") pod \"collect-profiles-29424045-lkk6q\" (UID: \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.287293 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-secret-volume\") pod \"collect-profiles-29424045-lkk6q\" (UID: \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.287386 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q9xm\" (UniqueName: \"kubernetes.io/projected/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-kube-api-access-7q9xm\") pod \"collect-profiles-29424045-lkk6q\" (UID: \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.288840 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-config-volume\") pod \"collect-profiles-29424045-lkk6q\" (UID: \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.294875 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-secret-volume\") pod \"collect-profiles-29424045-lkk6q\" (UID: \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.309996 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q9xm\" (UniqueName: \"kubernetes.io/projected/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-kube-api-access-7q9xm\") pod \"collect-profiles-29424045-lkk6q\" (UID: \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.501618 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:00 crc kubenswrapper[4629]: I1211 08:45:00.904923 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q"] Dec 11 08:45:01 crc kubenswrapper[4629]: I1211 08:45:01.796306 4629 generic.go:334] "Generic (PLEG): container finished" podID="c13206c5-13b9-4bf3-bdfd-7dd30d2ce643" containerID="eafab88ed89af90c86985f125c9b33a303b00063ec91c45f41c68421b6e95218" exitCode=0 Dec 11 08:45:01 crc kubenswrapper[4629]: I1211 08:45:01.796355 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" event={"ID":"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643","Type":"ContainerDied","Data":"eafab88ed89af90c86985f125c9b33a303b00063ec91c45f41c68421b6e95218"} Dec 11 08:45:01 crc kubenswrapper[4629]: I1211 08:45:01.796386 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" event={"ID":"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643","Type":"ContainerStarted","Data":"b460d84fa4712879736800c7c21226d4cd79ecc64791e46929d26bdef9c39e0f"} Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.112865 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.226359 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q9xm\" (UniqueName: \"kubernetes.io/projected/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-kube-api-access-7q9xm\") pod \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\" (UID: \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\") " Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.226455 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-secret-volume\") pod \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\" (UID: \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\") " Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.226491 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-config-volume\") pod \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\" (UID: \"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643\") " Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.227351 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-config-volume" (OuterVolumeSpecName: "config-volume") pod "c13206c5-13b9-4bf3-bdfd-7dd30d2ce643" (UID: "c13206c5-13b9-4bf3-bdfd-7dd30d2ce643"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.232318 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c13206c5-13b9-4bf3-bdfd-7dd30d2ce643" (UID: "c13206c5-13b9-4bf3-bdfd-7dd30d2ce643"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.233752 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-kube-api-access-7q9xm" (OuterVolumeSpecName: "kube-api-access-7q9xm") pod "c13206c5-13b9-4bf3-bdfd-7dd30d2ce643" (UID: "c13206c5-13b9-4bf3-bdfd-7dd30d2ce643"). InnerVolumeSpecName "kube-api-access-7q9xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.327796 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q9xm\" (UniqueName: \"kubernetes.io/projected/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-kube-api-access-7q9xm\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.327834 4629 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.327862 4629 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.808922 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" event={"ID":"c13206c5-13b9-4bf3-bdfd-7dd30d2ce643","Type":"ContainerDied","Data":"b460d84fa4712879736800c7c21226d4cd79ecc64791e46929d26bdef9c39e0f"} Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.809007 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b460d84fa4712879736800c7c21226d4cd79ecc64791e46929d26bdef9c39e0f" Dec 11 08:45:03 crc kubenswrapper[4629]: I1211 08:45:03.809013 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q" Dec 11 08:45:50 crc kubenswrapper[4629]: I1211 08:45:50.486062 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:45:50 crc kubenswrapper[4629]: I1211 08:45:50.486726 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:46:20 crc kubenswrapper[4629]: I1211 08:46:20.486304 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:46:20 crc kubenswrapper[4629]: I1211 08:46:20.490112 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:46:50 crc kubenswrapper[4629]: I1211 08:46:50.486599 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:46:50 crc kubenswrapper[4629]: I1211 08:46:50.487243 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:46:50 crc kubenswrapper[4629]: I1211 08:46:50.487319 4629 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:46:50 crc kubenswrapper[4629]: I1211 08:46:50.488142 4629 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"256bede4f7d04c63baf8ee22be7f6412368a168b18567d6d4b73acc6556f3040"} pod="openshift-machine-config-operator/machine-config-daemon-wbld9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:46:50 crc kubenswrapper[4629]: I1211 08:46:50.488230 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" containerID="cri-o://256bede4f7d04c63baf8ee22be7f6412368a168b18567d6d4b73acc6556f3040" gracePeriod=600 Dec 11 08:46:51 crc kubenswrapper[4629]: I1211 08:46:51.339774 4629 generic.go:334] "Generic (PLEG): container finished" podID="847b50b9-0093-404f-b0dd-63c814d8986b" containerID="256bede4f7d04c63baf8ee22be7f6412368a168b18567d6d4b73acc6556f3040" exitCode=0 Dec 11 08:46:51 crc kubenswrapper[4629]: I1211 08:46:51.339832 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerDied","Data":"256bede4f7d04c63baf8ee22be7f6412368a168b18567d6d4b73acc6556f3040"} Dec 11 08:46:51 crc kubenswrapper[4629]: I1211 08:46:51.340184 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"3a42a64840fd258d581b9d590c1921c344fcce2923f32019baf2139ce7bfec2f"} Dec 11 08:46:51 crc kubenswrapper[4629]: I1211 08:46:51.340206 4629 scope.go:117] "RemoveContainer" containerID="8ba210253304f5487d01920ea72fb77a38de3a8bfc04263290603cecc0511379" Dec 11 08:48:50 crc kubenswrapper[4629]: I1211 08:48:50.485527 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:48:50 crc kubenswrapper[4629]: I1211 08:48:50.486243 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:49:20 crc kubenswrapper[4629]: I1211 08:49:20.486440 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:49:20 crc kubenswrapper[4629]: I1211 08:49:20.487067 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.380807 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-f4zfd"] Dec 11 08:49:29 crc kubenswrapper[4629]: E1211 08:49:29.381714 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13206c5-13b9-4bf3-bdfd-7dd30d2ce643" containerName="collect-profiles" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.381730 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13206c5-13b9-4bf3-bdfd-7dd30d2ce643" containerName="collect-profiles" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.381891 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="c13206c5-13b9-4bf3-bdfd-7dd30d2ce643" containerName="collect-profiles" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.382409 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-f4zfd" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.387268 4629 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-rhxhz" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.387636 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.395718 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-f4zfd"] Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.396375 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.411123 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8td5b"] Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.411961 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-8td5b" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.415624 4629 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-sfz2h" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.426891 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nplz8"] Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.427773 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nplz8" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.430433 4629 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-2k8wp" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.443816 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8td5b"] Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.454181 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nplz8"] Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.578642 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thl28\" (UniqueName: \"kubernetes.io/projected/0fdee0b6-4a25-4531-bca0-adae607ea9c2-kube-api-access-thl28\") pod \"cert-manager-5b446d88c5-nplz8\" (UID: \"0fdee0b6-4a25-4531-bca0-adae607ea9c2\") " pod="cert-manager/cert-manager-5b446d88c5-nplz8" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.578924 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcklv\" (UniqueName: \"kubernetes.io/projected/5127d937-f916-4b39-bec4-79febeaeb994-kube-api-access-fcklv\") pod \"cert-manager-cainjector-7f985d654d-f4zfd\" (UID: \"5127d937-f916-4b39-bec4-79febeaeb994\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-f4zfd" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.579063 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjgz9\" (UniqueName: \"kubernetes.io/projected/17c0e2d5-d136-4e6b-a164-c43444ba024d-kube-api-access-zjgz9\") pod \"cert-manager-webhook-5655c58dd6-8td5b\" (UID: \"17c0e2d5-d136-4e6b-a164-c43444ba024d\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8td5b" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.679655 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjgz9\" (UniqueName: \"kubernetes.io/projected/17c0e2d5-d136-4e6b-a164-c43444ba024d-kube-api-access-zjgz9\") pod \"cert-manager-webhook-5655c58dd6-8td5b\" (UID: \"17c0e2d5-d136-4e6b-a164-c43444ba024d\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8td5b" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.679710 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thl28\" (UniqueName: \"kubernetes.io/projected/0fdee0b6-4a25-4531-bca0-adae607ea9c2-kube-api-access-thl28\") pod \"cert-manager-5b446d88c5-nplz8\" (UID: \"0fdee0b6-4a25-4531-bca0-adae607ea9c2\") " pod="cert-manager/cert-manager-5b446d88c5-nplz8" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.679746 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcklv\" (UniqueName: \"kubernetes.io/projected/5127d937-f916-4b39-bec4-79febeaeb994-kube-api-access-fcklv\") pod \"cert-manager-cainjector-7f985d654d-f4zfd\" (UID: \"5127d937-f916-4b39-bec4-79febeaeb994\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-f4zfd" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.700568 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thl28\" (UniqueName: \"kubernetes.io/projected/0fdee0b6-4a25-4531-bca0-adae607ea9c2-kube-api-access-thl28\") pod \"cert-manager-5b446d88c5-nplz8\" (UID: \"0fdee0b6-4a25-4531-bca0-adae607ea9c2\") " pod="cert-manager/cert-manager-5b446d88c5-nplz8" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.700867 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjgz9\" (UniqueName: \"kubernetes.io/projected/17c0e2d5-d136-4e6b-a164-c43444ba024d-kube-api-access-zjgz9\") pod \"cert-manager-webhook-5655c58dd6-8td5b\" (UID: \"17c0e2d5-d136-4e6b-a164-c43444ba024d\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8td5b" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.704020 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcklv\" (UniqueName: \"kubernetes.io/projected/5127d937-f916-4b39-bec4-79febeaeb994-kube-api-access-fcklv\") pod \"cert-manager-cainjector-7f985d654d-f4zfd\" (UID: \"5127d937-f916-4b39-bec4-79febeaeb994\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-f4zfd" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.708367 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-f4zfd" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.725478 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-8td5b" Dec 11 08:49:29 crc kubenswrapper[4629]: I1211 08:49:29.739699 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nplz8" Dec 11 08:49:30 crc kubenswrapper[4629]: I1211 08:49:30.163587 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-f4zfd"] Dec 11 08:49:30 crc kubenswrapper[4629]: I1211 08:49:30.173542 4629 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 08:49:30 crc kubenswrapper[4629]: I1211 08:49:30.259257 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8td5b"] Dec 11 08:49:30 crc kubenswrapper[4629]: I1211 08:49:30.269227 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nplz8"] Dec 11 08:49:30 crc kubenswrapper[4629]: W1211 08:49:30.276665 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fdee0b6_4a25_4531_bca0_adae607ea9c2.slice/crio-10ce0b703cf8256ecba9fc1454ce9d29039f770b22d7b54575970f2dad7bdc59 WatchSource:0}: Error finding container 10ce0b703cf8256ecba9fc1454ce9d29039f770b22d7b54575970f2dad7bdc59: Status 404 returned error can't find the container with id 10ce0b703cf8256ecba9fc1454ce9d29039f770b22d7b54575970f2dad7bdc59 Dec 11 08:49:31 crc kubenswrapper[4629]: I1211 08:49:31.151139 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nplz8" event={"ID":"0fdee0b6-4a25-4531-bca0-adae607ea9c2","Type":"ContainerStarted","Data":"10ce0b703cf8256ecba9fc1454ce9d29039f770b22d7b54575970f2dad7bdc59"} Dec 11 08:49:31 crc kubenswrapper[4629]: I1211 08:49:31.153679 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-f4zfd" event={"ID":"5127d937-f916-4b39-bec4-79febeaeb994","Type":"ContainerStarted","Data":"0abcaaba08533c30493374f935295363b01f6109d1a387b32b7ecf4d56306d6a"} Dec 11 08:49:31 crc kubenswrapper[4629]: I1211 08:49:31.155381 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-8td5b" event={"ID":"17c0e2d5-d136-4e6b-a164-c43444ba024d","Type":"ContainerStarted","Data":"7c6af82059f84351a9267c632c6e0dc2d8994f2ad472054c53632e7c71c988d3"} Dec 11 08:49:33 crc kubenswrapper[4629]: I1211 08:49:33.166920 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-8td5b" event={"ID":"17c0e2d5-d136-4e6b-a164-c43444ba024d","Type":"ContainerStarted","Data":"76e1030ac1f25a69eadcbcfef2bcb0da33bb3ea899999719e2d0d74607c9d493"} Dec 11 08:49:33 crc kubenswrapper[4629]: I1211 08:49:33.167555 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-8td5b" Dec 11 08:49:33 crc kubenswrapper[4629]: I1211 08:49:33.175075 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-f4zfd" event={"ID":"5127d937-f916-4b39-bec4-79febeaeb994","Type":"ContainerStarted","Data":"c7154060d0fb83019cfd3e08dee647c9d443bc5ed636625e8607a27c72ee4a8b"} Dec 11 08:49:33 crc kubenswrapper[4629]: I1211 08:49:33.182640 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-8td5b" podStartSLOduration=1.5576880800000001 podStartE2EDuration="4.182609784s" podCreationTimestamp="2025-12-11 08:49:29 +0000 UTC" firstStartedPulling="2025-12-11 08:49:30.271820619 +0000 UTC m=+758.337236227" lastFinishedPulling="2025-12-11 08:49:32.896742323 +0000 UTC m=+760.962157931" observedRunningTime="2025-12-11 08:49:33.180160678 +0000 UTC m=+761.245576296" watchObservedRunningTime="2025-12-11 08:49:33.182609784 +0000 UTC m=+761.248025392" Dec 11 08:49:33 crc kubenswrapper[4629]: I1211 08:49:33.200409 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-f4zfd" podStartSLOduration=1.473530487 podStartE2EDuration="4.200394734s" podCreationTimestamp="2025-12-11 08:49:29 +0000 UTC" firstStartedPulling="2025-12-11 08:49:30.173190928 +0000 UTC m=+758.238606536" lastFinishedPulling="2025-12-11 08:49:32.900055175 +0000 UTC m=+760.965470783" observedRunningTime="2025-12-11 08:49:33.197261627 +0000 UTC m=+761.262677225" watchObservedRunningTime="2025-12-11 08:49:33.200394734 +0000 UTC m=+761.265810342" Dec 11 08:49:35 crc kubenswrapper[4629]: I1211 08:49:35.185972 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nplz8" event={"ID":"0fdee0b6-4a25-4531-bca0-adae607ea9c2","Type":"ContainerStarted","Data":"fc1796ae328b2a6cfaafcf39b75d3a3255bf380bcfb36b9db7dd3a60aeaff615"} Dec 11 08:49:35 crc kubenswrapper[4629]: I1211 08:49:35.207000 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-nplz8" podStartSLOduration=2.324888868 podStartE2EDuration="6.206969973s" podCreationTimestamp="2025-12-11 08:49:29 +0000 UTC" firstStartedPulling="2025-12-11 08:49:30.279049843 +0000 UTC m=+758.344465451" lastFinishedPulling="2025-12-11 08:49:34.161130948 +0000 UTC m=+762.226546556" observedRunningTime="2025-12-11 08:49:35.206218411 +0000 UTC m=+763.271634019" watchObservedRunningTime="2025-12-11 08:49:35.206969973 +0000 UTC m=+763.272385591" Dec 11 08:49:35 crc kubenswrapper[4629]: I1211 08:49:35.906761 4629 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.356942 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lt499"] Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.357894 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovn-controller" containerID="cri-o://38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16" gracePeriod=30 Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.357928 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="nbdb" containerID="cri-o://209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66" gracePeriod=30 Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.358034 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="sbdb" containerID="cri-o://c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d" gracePeriod=30 Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.358128 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00" gracePeriod=30 Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.358143 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="northd" containerID="cri-o://531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd" gracePeriod=30 Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.358158 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="kube-rbac-proxy-node" containerID="cri-o://0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8" gracePeriod=30 Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.358175 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovn-acl-logging" containerID="cri-o://364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a" gracePeriod=30 Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.387440 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" containerID="cri-o://604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f" gracePeriod=30 Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.707790 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/3.log" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.711083 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovn-acl-logging/0.log" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.711530 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovn-controller/0.log" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.711942 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.732703 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-8td5b" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.779975 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vv95s"] Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.780401 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="sbdb" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780426 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="sbdb" Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.780442 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="kubecfg-setup" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780450 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="kubecfg-setup" Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.780468 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780478 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.780492 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="northd" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780500 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="northd" Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.780513 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="kube-rbac-proxy-node" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780521 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="kube-rbac-proxy-node" Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.780535 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780544 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.780555 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780562 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.780576 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780583 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.780592 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovn-acl-logging" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780600 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovn-acl-logging" Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.780614 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="nbdb" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780622 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="nbdb" Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.780631 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovn-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780639 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovn-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.780652 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="kube-rbac-proxy-ovn-metrics" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780659 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="kube-rbac-proxy-ovn-metrics" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780867 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovn-acl-logging" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780886 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780898 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780913 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="northd" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780928 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="sbdb" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780944 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="kube-rbac-proxy-node" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780956 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="kube-rbac-proxy-ovn-metrics" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780966 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780980 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="nbdb" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.780987 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovn-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: E1211 08:49:39.781210 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.781221 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.781412 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.781429 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerName="ovnkube-controller" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.785871 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898256 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-run-ovn-kubernetes\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898346 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-ovnkube-config\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898381 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-log-socket\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898383 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898404 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwkf6\" (UniqueName: \"kubernetes.io/projected/f89bfa10-15c0-40cd-b256-31a9934c855d-kube-api-access-rwkf6\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898471 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-kubelet\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898484 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-log-socket" (OuterVolumeSpecName: "log-socket") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898496 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-node-log\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898513 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-node-log" (OuterVolumeSpecName: "node-log") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898533 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898548 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-ovn\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898581 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898601 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-cni-bin\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898618 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-var-lib-openvswitch\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898624 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898645 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f89bfa10-15c0-40cd-b256-31a9934c855d-ovn-node-metrics-cert\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898646 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898661 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-cni-netd\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898664 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898682 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-openvswitch\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898687 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898699 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-slash\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898707 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898718 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-ovnkube-script-lib\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898744 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-systemd\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898776 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-etc-openvswitch\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898797 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-env-overrides\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898812 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-systemd-units\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898830 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-run-netns\") pod \"f89bfa10-15c0-40cd-b256-31a9934c855d\" (UID: \"f89bfa10-15c0-40cd-b256-31a9934c855d\") " Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.898999 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899028 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-run-systemd\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899061 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-slash\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899079 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-run-ovn-kubernetes\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899123 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-kubelet\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899144 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-cni-bin\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899164 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-etc-openvswitch\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899182 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f4dddc89-899a-44ef-8d73-204fde1692f7-ovn-node-metrics-cert\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899210 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-var-lib-openvswitch\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899224 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-node-log\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899243 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f4dddc89-899a-44ef-8d73-204fde1692f7-env-overrides\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899259 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899277 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-run-openvswitch\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899289 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-log-socket\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899304 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-run-ovn\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899344 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-cni-netd\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899357 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f4dddc89-899a-44ef-8d73-204fde1692f7-ovnkube-script-lib\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899375 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwzbb\" (UniqueName: \"kubernetes.io/projected/f4dddc89-899a-44ef-8d73-204fde1692f7-kube-api-access-mwzbb\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899396 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-run-netns\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899420 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-systemd-units\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899439 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f4dddc89-899a-44ef-8d73-204fde1692f7-ovnkube-config\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899512 4629 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899523 4629 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899532 4629 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899541 4629 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-log-socket\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899549 4629 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-node-log\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899557 4629 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899567 4629 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899575 4629 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899585 4629 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899621 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899939 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899969 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.899988 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.900401 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.900560 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-slash" (OuterVolumeSpecName: "host-slash") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.900957 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.903818 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f89bfa10-15c0-40cd-b256-31a9934c855d-kube-api-access-rwkf6" (OuterVolumeSpecName: "kube-api-access-rwkf6") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "kube-api-access-rwkf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.906547 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89bfa10-15c0-40cd-b256-31a9934c855d-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:49:39 crc kubenswrapper[4629]: I1211 08:49:39.912560 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "f89bfa10-15c0-40cd-b256-31a9934c855d" (UID: "f89bfa10-15c0-40cd-b256-31a9934c855d"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000279 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f4dddc89-899a-44ef-8d73-204fde1692f7-ovn-node-metrics-cert\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000336 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-var-lib-openvswitch\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000354 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-node-log\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000372 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f4dddc89-899a-44ef-8d73-204fde1692f7-env-overrides\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000388 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000405 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-run-openvswitch\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000418 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-log-socket\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000442 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-run-ovn\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000477 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-cni-netd\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000495 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f4dddc89-899a-44ef-8d73-204fde1692f7-ovnkube-script-lib\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000525 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwzbb\" (UniqueName: \"kubernetes.io/projected/f4dddc89-899a-44ef-8d73-204fde1692f7-kube-api-access-mwzbb\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000548 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-run-netns\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000573 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-systemd-units\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000600 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f4dddc89-899a-44ef-8d73-204fde1692f7-ovnkube-config\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000645 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-run-systemd\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000665 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-slash\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000681 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-run-ovn-kubernetes\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000715 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-kubelet\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000739 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-cni-bin\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000760 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-etc-openvswitch\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000807 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwkf6\" (UniqueName: \"kubernetes.io/projected/f89bfa10-15c0-40cd-b256-31a9934c855d-kube-api-access-rwkf6\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000823 4629 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000833 4629 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f89bfa10-15c0-40cd-b256-31a9934c855d-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000867 4629 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000879 4629 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-slash\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000889 4629 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000900 4629 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000911 4629 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.000960 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-run-netns\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001039 4629 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f89bfa10-15c0-40cd-b256-31a9934c855d-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001299 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f4dddc89-899a-44ef-8d73-204fde1692f7-env-overrides\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001374 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-node-log\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001417 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-cni-netd\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001448 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-kubelet\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001471 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-etc-openvswitch\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001489 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-systemd-units\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001562 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f4dddc89-899a-44ef-8d73-204fde1692f7-ovnkube-script-lib\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001605 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-run-ovn\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001628 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-cni-bin\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001647 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-var-lib-openvswitch\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001671 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-run-openvswitch\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001688 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001711 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-slash\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001728 4629 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001746 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-run-systemd\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001771 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-log-socket\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001796 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4dddc89-899a-44ef-8d73-204fde1692f7-host-run-ovn-kubernetes\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.001810 4629 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f89bfa10-15c0-40cd-b256-31a9934c855d-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.002170 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f4dddc89-899a-44ef-8d73-204fde1692f7-ovnkube-config\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.004193 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f4dddc89-899a-44ef-8d73-204fde1692f7-ovn-node-metrics-cert\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.018611 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwzbb\" (UniqueName: \"kubernetes.io/projected/f4dddc89-899a-44ef-8d73-204fde1692f7-kube-api-access-mwzbb\") pod \"ovnkube-node-vv95s\" (UID: \"f4dddc89-899a-44ef-8d73-204fde1692f7\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.101454 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.209254 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lvjmx_f5568b3d-2e9b-4038-bb16-4ad3d87435ee/kube-multus/2.log" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.209889 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lvjmx_f5568b3d-2e9b-4038-bb16-4ad3d87435ee/kube-multus/1.log" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.209919 4629 generic.go:334] "Generic (PLEG): container finished" podID="f5568b3d-2e9b-4038-bb16-4ad3d87435ee" containerID="a1c5e5255a2e972dcbb5f2f48978a4d9c479227775a20539ef2106ea7eb29d01" exitCode=2 Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.209962 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lvjmx" event={"ID":"f5568b3d-2e9b-4038-bb16-4ad3d87435ee","Type":"ContainerDied","Data":"a1c5e5255a2e972dcbb5f2f48978a4d9c479227775a20539ef2106ea7eb29d01"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.210230 4629 scope.go:117] "RemoveContainer" containerID="9c6c032f3ad347573ae69ef2fc9077703fd79c3d300ed8849a4cfd6c84ce5bf6" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.211681 4629 scope.go:117] "RemoveContainer" containerID="a1c5e5255a2e972dcbb5f2f48978a4d9c479227775a20539ef2106ea7eb29d01" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.213147 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovnkube-controller/3.log" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.215019 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovn-acl-logging/0.log" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.215650 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lt499_f89bfa10-15c0-40cd-b256-31a9934c855d/ovn-controller/0.log" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216123 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f" exitCode=0 Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216169 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d" exitCode=0 Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216179 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66" exitCode=0 Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216187 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd" exitCode=0 Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216194 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00" exitCode=0 Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216202 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8" exitCode=0 Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216209 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a" exitCode=143 Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216217 4629 generic.go:334] "Generic (PLEG): container finished" podID="f89bfa10-15c0-40cd-b256-31a9934c855d" containerID="38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16" exitCode=143 Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216284 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216329 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216342 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216351 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216360 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216371 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216400 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216413 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216420 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216425 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216431 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216437 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216442 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216447 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216452 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216400 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216476 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216588 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216610 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216620 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216626 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216631 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216637 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216642 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216647 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216652 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216657 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216661 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216668 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216678 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216686 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216691 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216696 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216700 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216706 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216710 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216716 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216723 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216727 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216734 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lt499" event={"ID":"f89bfa10-15c0-40cd-b256-31a9934c855d","Type":"ContainerDied","Data":"69a2b9f77a3183240a7feeadc3652a5779a90389c912b9dbd8b1b944421ae30c"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216742 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216748 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216753 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216759 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216763 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216768 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216774 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216779 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216784 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.216790 4629 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.231152 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" event={"ID":"f4dddc89-899a-44ef-8d73-204fde1692f7","Type":"ContainerStarted","Data":"19996a10d1a3223ea1b758cc22b02bb473428dd7d2fe5ea7c6fbbb222d0e003e"} Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.255534 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lt499"] Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.259345 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lt499"] Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.694949 4629 scope.go:117] "RemoveContainer" containerID="604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.714611 4629 scope.go:117] "RemoveContainer" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.753972 4629 scope.go:117] "RemoveContainer" containerID="c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.778434 4629 scope.go:117] "RemoveContainer" containerID="209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.796966 4629 scope.go:117] "RemoveContainer" containerID="531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.810971 4629 scope.go:117] "RemoveContainer" containerID="37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.822204 4629 scope.go:117] "RemoveContainer" containerID="0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.834831 4629 scope.go:117] "RemoveContainer" containerID="364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.848673 4629 scope.go:117] "RemoveContainer" containerID="38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.863010 4629 scope.go:117] "RemoveContainer" containerID="50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.875045 4629 scope.go:117] "RemoveContainer" containerID="604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f" Dec 11 08:49:40 crc kubenswrapper[4629]: E1211 08:49:40.875566 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f\": container with ID starting with 604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f not found: ID does not exist" containerID="604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.875619 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f"} err="failed to get container status \"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f\": rpc error: code = NotFound desc = could not find container \"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f\": container with ID starting with 604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.875649 4629 scope.go:117] "RemoveContainer" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" Dec 11 08:49:40 crc kubenswrapper[4629]: E1211 08:49:40.876056 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\": container with ID starting with 22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9 not found: ID does not exist" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.876093 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9"} err="failed to get container status \"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\": rpc error: code = NotFound desc = could not find container \"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\": container with ID starting with 22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.876173 4629 scope.go:117] "RemoveContainer" containerID="c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d" Dec 11 08:49:40 crc kubenswrapper[4629]: E1211 08:49:40.876468 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\": container with ID starting with c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d not found: ID does not exist" containerID="c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.876496 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d"} err="failed to get container status \"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\": rpc error: code = NotFound desc = could not find container \"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\": container with ID starting with c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.876514 4629 scope.go:117] "RemoveContainer" containerID="209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66" Dec 11 08:49:40 crc kubenswrapper[4629]: E1211 08:49:40.876737 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\": container with ID starting with 209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66 not found: ID does not exist" containerID="209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.876762 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66"} err="failed to get container status \"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\": rpc error: code = NotFound desc = could not find container \"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\": container with ID starting with 209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.876779 4629 scope.go:117] "RemoveContainer" containerID="531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd" Dec 11 08:49:40 crc kubenswrapper[4629]: E1211 08:49:40.878055 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\": container with ID starting with 531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd not found: ID does not exist" containerID="531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.878175 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd"} err="failed to get container status \"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\": rpc error: code = NotFound desc = could not find container \"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\": container with ID starting with 531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.878286 4629 scope.go:117] "RemoveContainer" containerID="37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00" Dec 11 08:49:40 crc kubenswrapper[4629]: E1211 08:49:40.878666 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\": container with ID starting with 37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00 not found: ID does not exist" containerID="37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.878698 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00"} err="failed to get container status \"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\": rpc error: code = NotFound desc = could not find container \"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\": container with ID starting with 37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.878719 4629 scope.go:117] "RemoveContainer" containerID="0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8" Dec 11 08:49:40 crc kubenswrapper[4629]: E1211 08:49:40.879034 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\": container with ID starting with 0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8 not found: ID does not exist" containerID="0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.879061 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8"} err="failed to get container status \"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\": rpc error: code = NotFound desc = could not find container \"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\": container with ID starting with 0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.879080 4629 scope.go:117] "RemoveContainer" containerID="364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a" Dec 11 08:49:40 crc kubenswrapper[4629]: E1211 08:49:40.879419 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\": container with ID starting with 364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a not found: ID does not exist" containerID="364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.879447 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a"} err="failed to get container status \"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\": rpc error: code = NotFound desc = could not find container \"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\": container with ID starting with 364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.879464 4629 scope.go:117] "RemoveContainer" containerID="38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16" Dec 11 08:49:40 crc kubenswrapper[4629]: E1211 08:49:40.880111 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\": container with ID starting with 38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16 not found: ID does not exist" containerID="38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.880144 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16"} err="failed to get container status \"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\": rpc error: code = NotFound desc = could not find container \"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\": container with ID starting with 38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.880163 4629 scope.go:117] "RemoveContainer" containerID="50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272" Dec 11 08:49:40 crc kubenswrapper[4629]: E1211 08:49:40.880424 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\": container with ID starting with 50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272 not found: ID does not exist" containerID="50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.880452 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272"} err="failed to get container status \"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\": rpc error: code = NotFound desc = could not find container \"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\": container with ID starting with 50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.880471 4629 scope.go:117] "RemoveContainer" containerID="604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.880722 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f"} err="failed to get container status \"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f\": rpc error: code = NotFound desc = could not find container \"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f\": container with ID starting with 604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.880743 4629 scope.go:117] "RemoveContainer" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.881035 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9"} err="failed to get container status \"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\": rpc error: code = NotFound desc = could not find container \"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\": container with ID starting with 22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.881067 4629 scope.go:117] "RemoveContainer" containerID="c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.881271 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d"} err="failed to get container status \"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\": rpc error: code = NotFound desc = could not find container \"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\": container with ID starting with c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.881320 4629 scope.go:117] "RemoveContainer" containerID="209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.881595 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66"} err="failed to get container status \"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\": rpc error: code = NotFound desc = could not find container \"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\": container with ID starting with 209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.881647 4629 scope.go:117] "RemoveContainer" containerID="531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.882589 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd"} err="failed to get container status \"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\": rpc error: code = NotFound desc = could not find container \"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\": container with ID starting with 531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.882618 4629 scope.go:117] "RemoveContainer" containerID="37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.882882 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00"} err="failed to get container status \"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\": rpc error: code = NotFound desc = could not find container \"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\": container with ID starting with 37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.882907 4629 scope.go:117] "RemoveContainer" containerID="0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.883167 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8"} err="failed to get container status \"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\": rpc error: code = NotFound desc = could not find container \"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\": container with ID starting with 0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.883260 4629 scope.go:117] "RemoveContainer" containerID="364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.883603 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a"} err="failed to get container status \"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\": rpc error: code = NotFound desc = could not find container \"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\": container with ID starting with 364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.883629 4629 scope.go:117] "RemoveContainer" containerID="38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.883880 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16"} err="failed to get container status \"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\": rpc error: code = NotFound desc = could not find container \"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\": container with ID starting with 38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.883907 4629 scope.go:117] "RemoveContainer" containerID="50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.884131 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272"} err="failed to get container status \"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\": rpc error: code = NotFound desc = could not find container \"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\": container with ID starting with 50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.884155 4629 scope.go:117] "RemoveContainer" containerID="604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.884440 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f"} err="failed to get container status \"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f\": rpc error: code = NotFound desc = could not find container \"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f\": container with ID starting with 604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.884539 4629 scope.go:117] "RemoveContainer" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.884967 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9"} err="failed to get container status \"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\": rpc error: code = NotFound desc = could not find container \"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\": container with ID starting with 22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.884994 4629 scope.go:117] "RemoveContainer" containerID="c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.885192 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d"} err="failed to get container status \"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\": rpc error: code = NotFound desc = could not find container \"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\": container with ID starting with c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.885221 4629 scope.go:117] "RemoveContainer" containerID="209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.885464 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66"} err="failed to get container status \"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\": rpc error: code = NotFound desc = could not find container \"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\": container with ID starting with 209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.885494 4629 scope.go:117] "RemoveContainer" containerID="531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.885668 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd"} err="failed to get container status \"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\": rpc error: code = NotFound desc = could not find container \"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\": container with ID starting with 531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.885696 4629 scope.go:117] "RemoveContainer" containerID="37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.885972 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00"} err="failed to get container status \"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\": rpc error: code = NotFound desc = could not find container \"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\": container with ID starting with 37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.885994 4629 scope.go:117] "RemoveContainer" containerID="0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.886314 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8"} err="failed to get container status \"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\": rpc error: code = NotFound desc = could not find container \"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\": container with ID starting with 0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.886334 4629 scope.go:117] "RemoveContainer" containerID="364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.886540 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a"} err="failed to get container status \"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\": rpc error: code = NotFound desc = could not find container \"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\": container with ID starting with 364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.886558 4629 scope.go:117] "RemoveContainer" containerID="38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.886753 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16"} err="failed to get container status \"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\": rpc error: code = NotFound desc = could not find container \"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\": container with ID starting with 38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.886775 4629 scope.go:117] "RemoveContainer" containerID="50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.886986 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272"} err="failed to get container status \"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\": rpc error: code = NotFound desc = could not find container \"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\": container with ID starting with 50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.887005 4629 scope.go:117] "RemoveContainer" containerID="604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.887174 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f"} err="failed to get container status \"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f\": rpc error: code = NotFound desc = could not find container \"604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f\": container with ID starting with 604e88f5b4f0925300d38182caa7a345020827d21cb03fe220484c77c5a8388f not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.887193 4629 scope.go:117] "RemoveContainer" containerID="22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.887388 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9"} err="failed to get container status \"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\": rpc error: code = NotFound desc = could not find container \"22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9\": container with ID starting with 22cd49803d21faa590febdcb3befa1f05827021358f26540d856c3710c9a5fc9 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.887405 4629 scope.go:117] "RemoveContainer" containerID="c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.887620 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d"} err="failed to get container status \"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\": rpc error: code = NotFound desc = could not find container \"c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d\": container with ID starting with c2c56d7073caeb3ecb73eb3e7192e001aaaed6f43e6e6b8e043689c42ed5280d not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.887641 4629 scope.go:117] "RemoveContainer" containerID="209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.887896 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66"} err="failed to get container status \"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\": rpc error: code = NotFound desc = could not find container \"209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66\": container with ID starting with 209c362517d2707937e53c5c5b455f4f9a079799040be610d07c02af3993cd66 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.887916 4629 scope.go:117] "RemoveContainer" containerID="531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.888609 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd"} err="failed to get container status \"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\": rpc error: code = NotFound desc = could not find container \"531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd\": container with ID starting with 531a76e228a9b3790884ef3e319a3b011ff61377c122be69b3f533aa7e8526dd not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.888633 4629 scope.go:117] "RemoveContainer" containerID="37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.888833 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00"} err="failed to get container status \"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\": rpc error: code = NotFound desc = could not find container \"37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00\": container with ID starting with 37f6625d894c4c74c754da8dbcb5ad2e7d415b97521fae33eac097abe5f71d00 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.888875 4629 scope.go:117] "RemoveContainer" containerID="0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.889349 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8"} err="failed to get container status \"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\": rpc error: code = NotFound desc = could not find container \"0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8\": container with ID starting with 0ca6a1eccea44095d7fbfbac6da3ac4c3077bcecaecb3dd6c7409a76f76e38f8 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.889378 4629 scope.go:117] "RemoveContainer" containerID="364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.889592 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a"} err="failed to get container status \"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\": rpc error: code = NotFound desc = could not find container \"364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a\": container with ID starting with 364c39a5ece11cb17463bceb67b3a76043ed6e49a079a53f9b017b6e15a6760a not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.889618 4629 scope.go:117] "RemoveContainer" containerID="38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.889907 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16"} err="failed to get container status \"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\": rpc error: code = NotFound desc = could not find container \"38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16\": container with ID starting with 38f73f7eeb799a7f612f1044487b923c99369a5a00797db5e651726620b5ed16 not found: ID does not exist" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.889961 4629 scope.go:117] "RemoveContainer" containerID="50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272" Dec 11 08:49:40 crc kubenswrapper[4629]: I1211 08:49:40.890195 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272"} err="failed to get container status \"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\": rpc error: code = NotFound desc = could not find container \"50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272\": container with ID starting with 50550a004de073515c6f89fdec12aea11e5fad929bafd735d5a3207e37775272 not found: ID does not exist" Dec 11 08:49:41 crc kubenswrapper[4629]: I1211 08:49:41.242550 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lvjmx_f5568b3d-2e9b-4038-bb16-4ad3d87435ee/kube-multus/2.log" Dec 11 08:49:41 crc kubenswrapper[4629]: I1211 08:49:41.242926 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lvjmx" event={"ID":"f5568b3d-2e9b-4038-bb16-4ad3d87435ee","Type":"ContainerStarted","Data":"1b5a19a2384639b81e92ab11e5bd9325eadcc7cfbaa0c5e0d70a6ca927ef09d8"} Dec 11 08:49:41 crc kubenswrapper[4629]: I1211 08:49:41.247134 4629 generic.go:334] "Generic (PLEG): container finished" podID="f4dddc89-899a-44ef-8d73-204fde1692f7" containerID="100a6228a07970c552cf0ca306d148646a9796c961d87faa731c1f1bff9782ed" exitCode=0 Dec 11 08:49:41 crc kubenswrapper[4629]: I1211 08:49:41.247307 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" event={"ID":"f4dddc89-899a-44ef-8d73-204fde1692f7","Type":"ContainerDied","Data":"100a6228a07970c552cf0ca306d148646a9796c961d87faa731c1f1bff9782ed"} Dec 11 08:49:42 crc kubenswrapper[4629]: I1211 08:49:42.205483 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f89bfa10-15c0-40cd-b256-31a9934c855d" path="/var/lib/kubelet/pods/f89bfa10-15c0-40cd-b256-31a9934c855d/volumes" Dec 11 08:49:42 crc kubenswrapper[4629]: I1211 08:49:42.255705 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" event={"ID":"f4dddc89-899a-44ef-8d73-204fde1692f7","Type":"ContainerStarted","Data":"c92a8918b67441e78d26012218114ae531aa6b47c75bd02daccbcd247f01bce2"} Dec 11 08:49:42 crc kubenswrapper[4629]: I1211 08:49:42.255761 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" event={"ID":"f4dddc89-899a-44ef-8d73-204fde1692f7","Type":"ContainerStarted","Data":"fa0ccc9c36117520cc61fa919c65a73fda0869532fde77c35beed2c38b805bd1"} Dec 11 08:49:42 crc kubenswrapper[4629]: I1211 08:49:42.255780 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" event={"ID":"f4dddc89-899a-44ef-8d73-204fde1692f7","Type":"ContainerStarted","Data":"47989be3a4a570fd3a84873847a401506e6499712a4fbf6e2677a4aafb5c1058"} Dec 11 08:49:42 crc kubenswrapper[4629]: I1211 08:49:42.255792 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" event={"ID":"f4dddc89-899a-44ef-8d73-204fde1692f7","Type":"ContainerStarted","Data":"2f9c72166e6895004ee9b98e3ef62a3a4ad3ca761253542b39ee6655f146dd1c"} Dec 11 08:49:42 crc kubenswrapper[4629]: I1211 08:49:42.255801 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" event={"ID":"f4dddc89-899a-44ef-8d73-204fde1692f7","Type":"ContainerStarted","Data":"91412da891a3eb70b7a706c4340f503c6302c47918a37bd827c395c54d0299dc"} Dec 11 08:49:42 crc kubenswrapper[4629]: I1211 08:49:42.255809 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" event={"ID":"f4dddc89-899a-44ef-8d73-204fde1692f7","Type":"ContainerStarted","Data":"45f695b8ff183d868dbcd62998aa021492dd61e166a971a192048f330635a9ea"} Dec 11 08:49:44 crc kubenswrapper[4629]: I1211 08:49:44.271481 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" event={"ID":"f4dddc89-899a-44ef-8d73-204fde1692f7","Type":"ContainerStarted","Data":"0998a5d61d5f91a756c95ac06069c1a7c83d9bff8d35be84b3583c1204ac1ab5"} Dec 11 08:49:47 crc kubenswrapper[4629]: I1211 08:49:47.306689 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" event={"ID":"f4dddc89-899a-44ef-8d73-204fde1692f7","Type":"ContainerStarted","Data":"a364de4041bc29bf4e5b3b84465408e8b4be1abc3247a506847c177660074cb1"} Dec 11 08:49:47 crc kubenswrapper[4629]: I1211 08:49:47.308150 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:47 crc kubenswrapper[4629]: I1211 08:49:47.308175 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:47 crc kubenswrapper[4629]: I1211 08:49:47.308186 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:47 crc kubenswrapper[4629]: I1211 08:49:47.345514 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" podStartSLOduration=8.345489745 podStartE2EDuration="8.345489745s" podCreationTimestamp="2025-12-11 08:49:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:49:47.342515173 +0000 UTC m=+775.407930781" watchObservedRunningTime="2025-12-11 08:49:47.345489745 +0000 UTC m=+775.410905353" Dec 11 08:49:47 crc kubenswrapper[4629]: I1211 08:49:47.348477 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:47 crc kubenswrapper[4629]: I1211 08:49:47.348541 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:49:50 crc kubenswrapper[4629]: I1211 08:49:50.485549 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:49:50 crc kubenswrapper[4629]: I1211 08:49:50.486007 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:49:50 crc kubenswrapper[4629]: I1211 08:49:50.486064 4629 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:49:50 crc kubenswrapper[4629]: I1211 08:49:50.486839 4629 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3a42a64840fd258d581b9d590c1921c344fcce2923f32019baf2139ce7bfec2f"} pod="openshift-machine-config-operator/machine-config-daemon-wbld9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:49:50 crc kubenswrapper[4629]: I1211 08:49:50.486935 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" containerID="cri-o://3a42a64840fd258d581b9d590c1921c344fcce2923f32019baf2139ce7bfec2f" gracePeriod=600 Dec 11 08:49:51 crc kubenswrapper[4629]: I1211 08:49:51.344930 4629 generic.go:334] "Generic (PLEG): container finished" podID="847b50b9-0093-404f-b0dd-63c814d8986b" containerID="3a42a64840fd258d581b9d590c1921c344fcce2923f32019baf2139ce7bfec2f" exitCode=0 Dec 11 08:49:51 crc kubenswrapper[4629]: I1211 08:49:51.344999 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerDied","Data":"3a42a64840fd258d581b9d590c1921c344fcce2923f32019baf2139ce7bfec2f"} Dec 11 08:49:51 crc kubenswrapper[4629]: I1211 08:49:51.345301 4629 scope.go:117] "RemoveContainer" containerID="256bede4f7d04c63baf8ee22be7f6412368a168b18567d6d4b73acc6556f3040" Dec 11 08:49:52 crc kubenswrapper[4629]: I1211 08:49:52.353063 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"3478d1305c4015a5af7c390010bcd373d9a5066e2359bd5ed303da8f30af223b"} Dec 11 08:50:10 crc kubenswrapper[4629]: I1211 08:50:10.128232 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vv95s" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.062977 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs"] Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.064653 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.068730 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.080314 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs"] Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.195889 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs\" (UID: \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.195967 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp2kj\" (UniqueName: \"kubernetes.io/projected/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-kube-api-access-fp2kj\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs\" (UID: \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.196052 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs\" (UID: \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.296760 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp2kj\" (UniqueName: \"kubernetes.io/projected/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-kube-api-access-fp2kj\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs\" (UID: \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.296845 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs\" (UID: \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.296941 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs\" (UID: \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.297815 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs\" (UID: \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.298608 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs\" (UID: \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.315579 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp2kj\" (UniqueName: \"kubernetes.io/projected/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-kube-api-access-fp2kj\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs\" (UID: \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.390325 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:22 crc kubenswrapper[4629]: I1211 08:50:22.919491 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs"] Dec 11 08:50:23 crc kubenswrapper[4629]: I1211 08:50:23.629539 4629 generic.go:334] "Generic (PLEG): container finished" podID="a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" containerID="47b11c59d504e0e757ca73c1e8514371af89682e199bb1bab0a93b1cfcdd40c1" exitCode=0 Dec 11 08:50:23 crc kubenswrapper[4629]: I1211 08:50:23.629663 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" event={"ID":"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9","Type":"ContainerDied","Data":"47b11c59d504e0e757ca73c1e8514371af89682e199bb1bab0a93b1cfcdd40c1"} Dec 11 08:50:23 crc kubenswrapper[4629]: I1211 08:50:23.630026 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" event={"ID":"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9","Type":"ContainerStarted","Data":"459e455b8ff255fc37bb14ab84998187d9c21259888ace7bbbe51d483173367e"} Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.419759 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z2vwh"] Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.421750 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.427813 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z2vwh"] Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.526465 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md2l4\" (UniqueName: \"kubernetes.io/projected/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-kube-api-access-md2l4\") pod \"redhat-operators-z2vwh\" (UID: \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\") " pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.526547 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-utilities\") pod \"redhat-operators-z2vwh\" (UID: \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\") " pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.526604 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-catalog-content\") pod \"redhat-operators-z2vwh\" (UID: \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\") " pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.627955 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md2l4\" (UniqueName: \"kubernetes.io/projected/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-kube-api-access-md2l4\") pod \"redhat-operators-z2vwh\" (UID: \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\") " pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.628360 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-utilities\") pod \"redhat-operators-z2vwh\" (UID: \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\") " pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.628970 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-utilities\") pod \"redhat-operators-z2vwh\" (UID: \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\") " pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.629050 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-catalog-content\") pod \"redhat-operators-z2vwh\" (UID: \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\") " pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.629377 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-catalog-content\") pod \"redhat-operators-z2vwh\" (UID: \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\") " pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.649106 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md2l4\" (UniqueName: \"kubernetes.io/projected/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-kube-api-access-md2l4\") pod \"redhat-operators-z2vwh\" (UID: \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\") " pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.740762 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:24 crc kubenswrapper[4629]: I1211 08:50:24.960327 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z2vwh"] Dec 11 08:50:25 crc kubenswrapper[4629]: I1211 08:50:25.642542 4629 generic.go:334] "Generic (PLEG): container finished" podID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" containerID="4e9d2c90cf7f173d71fe44f80b385e666ad5bc39d1f02a984263f9566b151aec" exitCode=0 Dec 11 08:50:25 crc kubenswrapper[4629]: I1211 08:50:25.642975 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2vwh" event={"ID":"5fe1b4ed-87d1-4802-8350-59e3ed522eeb","Type":"ContainerDied","Data":"4e9d2c90cf7f173d71fe44f80b385e666ad5bc39d1f02a984263f9566b151aec"} Dec 11 08:50:25 crc kubenswrapper[4629]: I1211 08:50:25.643026 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2vwh" event={"ID":"5fe1b4ed-87d1-4802-8350-59e3ed522eeb","Type":"ContainerStarted","Data":"c54a69aafca0ff7cac1e480daf7166d752311d8c3aef05ea03bd0aecfeedab08"} Dec 11 08:50:25 crc kubenswrapper[4629]: I1211 08:50:25.647711 4629 generic.go:334] "Generic (PLEG): container finished" podID="a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" containerID="c809ed885ef7a8c66f9023515e00435d2ff4b9671d4d8bd0ef6bde67d9ca0806" exitCode=0 Dec 11 08:50:25 crc kubenswrapper[4629]: I1211 08:50:25.647755 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" event={"ID":"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9","Type":"ContainerDied","Data":"c809ed885ef7a8c66f9023515e00435d2ff4b9671d4d8bd0ef6bde67d9ca0806"} Dec 11 08:50:26 crc kubenswrapper[4629]: I1211 08:50:26.654368 4629 generic.go:334] "Generic (PLEG): container finished" podID="a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" containerID="910f9f831d5ddb9fbe7494562242b4aa78b3fe5e4559b78eadddf5f8062fddef" exitCode=0 Dec 11 08:50:26 crc kubenswrapper[4629]: I1211 08:50:26.654401 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" event={"ID":"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9","Type":"ContainerDied","Data":"910f9f831d5ddb9fbe7494562242b4aa78b3fe5e4559b78eadddf5f8062fddef"} Dec 11 08:50:27 crc kubenswrapper[4629]: I1211 08:50:27.663169 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2vwh" event={"ID":"5fe1b4ed-87d1-4802-8350-59e3ed522eeb","Type":"ContainerStarted","Data":"393487a3acae4fed319c25a331cce5684717d880714a684a15b5c27c42ac7c66"} Dec 11 08:50:28 crc kubenswrapper[4629]: I1211 08:50:28.674005 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" event={"ID":"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9","Type":"ContainerDied","Data":"459e455b8ff255fc37bb14ab84998187d9c21259888ace7bbbe51d483173367e"} Dec 11 08:50:28 crc kubenswrapper[4629]: I1211 08:50:28.674349 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="459e455b8ff255fc37bb14ab84998187d9c21259888ace7bbbe51d483173367e" Dec 11 08:50:28 crc kubenswrapper[4629]: I1211 08:50:28.709208 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:28 crc kubenswrapper[4629]: I1211 08:50:28.881741 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp2kj\" (UniqueName: \"kubernetes.io/projected/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-kube-api-access-fp2kj\") pod \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\" (UID: \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\") " Dec 11 08:50:28 crc kubenswrapper[4629]: I1211 08:50:28.881794 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-util\") pod \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\" (UID: \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\") " Dec 11 08:50:28 crc kubenswrapper[4629]: I1211 08:50:28.881935 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-bundle\") pod \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\" (UID: \"a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9\") " Dec 11 08:50:28 crc kubenswrapper[4629]: I1211 08:50:28.882672 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-bundle" (OuterVolumeSpecName: "bundle") pod "a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" (UID: "a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:50:28 crc kubenswrapper[4629]: I1211 08:50:28.900394 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-util" (OuterVolumeSpecName: "util") pod "a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" (UID: "a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:50:28 crc kubenswrapper[4629]: I1211 08:50:28.983686 4629 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:50:28 crc kubenswrapper[4629]: I1211 08:50:28.983735 4629 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-util\") on node \"crc\" DevicePath \"\"" Dec 11 08:50:28 crc kubenswrapper[4629]: I1211 08:50:28.996397 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-kube-api-access-fp2kj" (OuterVolumeSpecName: "kube-api-access-fp2kj") pod "a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" (UID: "a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9"). InnerVolumeSpecName "kube-api-access-fp2kj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:50:29 crc kubenswrapper[4629]: I1211 08:50:29.084571 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp2kj\" (UniqueName: \"kubernetes.io/projected/a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9-kube-api-access-fp2kj\") on node \"crc\" DevicePath \"\"" Dec 11 08:50:29 crc kubenswrapper[4629]: I1211 08:50:29.678881 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs" Dec 11 08:50:30 crc kubenswrapper[4629]: I1211 08:50:30.696030 4629 generic.go:334] "Generic (PLEG): container finished" podID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" containerID="393487a3acae4fed319c25a331cce5684717d880714a684a15b5c27c42ac7c66" exitCode=0 Dec 11 08:50:30 crc kubenswrapper[4629]: I1211 08:50:30.696076 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2vwh" event={"ID":"5fe1b4ed-87d1-4802-8350-59e3ed522eeb","Type":"ContainerDied","Data":"393487a3acae4fed319c25a331cce5684717d880714a684a15b5c27c42ac7c66"} Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.707383 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2vwh" event={"ID":"5fe1b4ed-87d1-4802-8350-59e3ed522eeb","Type":"ContainerStarted","Data":"eafa7c9efe588b9ed4df2e19d4ccae2731febcb9ef5c3e9daa96625fc1589cc4"} Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.743545 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z2vwh" podStartSLOduration=2.040322336 podStartE2EDuration="7.743513624s" podCreationTimestamp="2025-12-11 08:50:24 +0000 UTC" firstStartedPulling="2025-12-11 08:50:25.645419861 +0000 UTC m=+813.710835469" lastFinishedPulling="2025-12-11 08:50:31.348611149 +0000 UTC m=+819.414026757" observedRunningTime="2025-12-11 08:50:31.739663062 +0000 UTC m=+819.805078700" watchObservedRunningTime="2025-12-11 08:50:31.743513624 +0000 UTC m=+819.808929222" Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.816477 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-kbnnp"] Dec 11 08:50:31 crc kubenswrapper[4629]: E1211 08:50:31.816690 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" containerName="pull" Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.816701 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" containerName="pull" Dec 11 08:50:31 crc kubenswrapper[4629]: E1211 08:50:31.816713 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" containerName="util" Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.816719 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" containerName="util" Dec 11 08:50:31 crc kubenswrapper[4629]: E1211 08:50:31.816727 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" containerName="extract" Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.816734 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" containerName="extract" Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.816822 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9" containerName="extract" Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.817235 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-kbnnp" Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.819989 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.820044 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-f5rx9" Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.820110 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.838174 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-kbnnp"] Dec 11 08:50:31 crc kubenswrapper[4629]: I1211 08:50:31.983942 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttxth\" (UniqueName: \"kubernetes.io/projected/e6f751dc-5c3e-4426-a89d-a38a936092e3-kube-api-access-ttxth\") pod \"nmstate-operator-6769fb99d-kbnnp\" (UID: \"e6f751dc-5c3e-4426-a89d-a38a936092e3\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-kbnnp" Dec 11 08:50:32 crc kubenswrapper[4629]: I1211 08:50:32.085239 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttxth\" (UniqueName: \"kubernetes.io/projected/e6f751dc-5c3e-4426-a89d-a38a936092e3-kube-api-access-ttxth\") pod \"nmstate-operator-6769fb99d-kbnnp\" (UID: \"e6f751dc-5c3e-4426-a89d-a38a936092e3\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-kbnnp" Dec 11 08:50:32 crc kubenswrapper[4629]: I1211 08:50:32.111435 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttxth\" (UniqueName: \"kubernetes.io/projected/e6f751dc-5c3e-4426-a89d-a38a936092e3-kube-api-access-ttxth\") pod \"nmstate-operator-6769fb99d-kbnnp\" (UID: \"e6f751dc-5c3e-4426-a89d-a38a936092e3\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-kbnnp" Dec 11 08:50:32 crc kubenswrapper[4629]: I1211 08:50:32.132916 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-kbnnp" Dec 11 08:50:32 crc kubenswrapper[4629]: I1211 08:50:32.572664 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-kbnnp"] Dec 11 08:50:32 crc kubenswrapper[4629]: W1211 08:50:32.578407 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6f751dc_5c3e_4426_a89d_a38a936092e3.slice/crio-a588b03055bced5b028cc8f3ea015a9a6f39956729901c9bf0d0cfa3ce568560 WatchSource:0}: Error finding container a588b03055bced5b028cc8f3ea015a9a6f39956729901c9bf0d0cfa3ce568560: Status 404 returned error can't find the container with id a588b03055bced5b028cc8f3ea015a9a6f39956729901c9bf0d0cfa3ce568560 Dec 11 08:50:32 crc kubenswrapper[4629]: I1211 08:50:32.712885 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-kbnnp" event={"ID":"e6f751dc-5c3e-4426-a89d-a38a936092e3","Type":"ContainerStarted","Data":"a588b03055bced5b028cc8f3ea015a9a6f39956729901c9bf0d0cfa3ce568560"} Dec 11 08:50:34 crc kubenswrapper[4629]: I1211 08:50:34.741878 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:34 crc kubenswrapper[4629]: I1211 08:50:34.742237 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:35 crc kubenswrapper[4629]: I1211 08:50:35.893288 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z2vwh" podUID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" containerName="registry-server" probeResult="failure" output=< Dec 11 08:50:35 crc kubenswrapper[4629]: timeout: failed to connect service ":50051" within 1s Dec 11 08:50:35 crc kubenswrapper[4629]: > Dec 11 08:50:36 crc kubenswrapper[4629]: I1211 08:50:36.748959 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-kbnnp" event={"ID":"e6f751dc-5c3e-4426-a89d-a38a936092e3","Type":"ContainerStarted","Data":"78d1a318917239018f255eef6fc0bcf018f243c1bc9bdd730d023ab7793d4830"} Dec 11 08:50:36 crc kubenswrapper[4629]: I1211 08:50:36.767150 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-6769fb99d-kbnnp" podStartSLOduration=2.190670755 podStartE2EDuration="5.76712804s" podCreationTimestamp="2025-12-11 08:50:31 +0000 UTC" firstStartedPulling="2025-12-11 08:50:32.580698391 +0000 UTC m=+820.646113999" lastFinishedPulling="2025-12-11 08:50:36.157155676 +0000 UTC m=+824.222571284" observedRunningTime="2025-12-11 08:50:36.762767573 +0000 UTC m=+824.828183201" watchObservedRunningTime="2025-12-11 08:50:36.76712804 +0000 UTC m=+824.832543658" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.243741 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-f64ct"] Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.245450 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-f64ct" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.247377 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-r78dj" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.265693 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs"] Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.266542 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.269085 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-f64ct"] Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.276096 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs"] Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.276329 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.283955 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-t7j8b"] Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.284871 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.383034 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4c5aa8ca-24cf-4ac8-937d-349e4d15eae2-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-vtqqs\" (UID: \"4c5aa8ca-24cf-4ac8-937d-349e4d15eae2\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.383131 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdlpc\" (UniqueName: \"kubernetes.io/projected/4c5aa8ca-24cf-4ac8-937d-349e4d15eae2-kube-api-access-fdlpc\") pod \"nmstate-webhook-f8fb84555-vtqqs\" (UID: \"4c5aa8ca-24cf-4ac8-937d-349e4d15eae2\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.383156 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frcsl\" (UniqueName: \"kubernetes.io/projected/39b77803-5bb8-436f-8ca6-bc4297ea81e8-kube-api-access-frcsl\") pod \"nmstate-metrics-7f7f7578db-f64ct\" (UID: \"39b77803-5bb8-436f-8ca6-bc4297ea81e8\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-f64ct" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.457455 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc"] Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.458147 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.462025 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.462592 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-jzgtx" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.464205 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.484571 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d89bee79-f8ff-4b46-98d0-b76671baf508-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-nm6nc\" (UID: \"d89bee79-f8ff-4b46-98d0-b76671baf508\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.484651 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4c5aa8ca-24cf-4ac8-937d-349e4d15eae2-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-vtqqs\" (UID: \"4c5aa8ca-24cf-4ac8-937d-349e4d15eae2\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.484685 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmcr6\" (UniqueName: \"kubernetes.io/projected/d89bee79-f8ff-4b46-98d0-b76671baf508-kube-api-access-dmcr6\") pod \"nmstate-console-plugin-6ff7998486-nm6nc\" (UID: \"d89bee79-f8ff-4b46-98d0-b76671baf508\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.484709 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdlpc\" (UniqueName: \"kubernetes.io/projected/4c5aa8ca-24cf-4ac8-937d-349e4d15eae2-kube-api-access-fdlpc\") pod \"nmstate-webhook-f8fb84555-vtqqs\" (UID: \"4c5aa8ca-24cf-4ac8-937d-349e4d15eae2\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.484735 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frcsl\" (UniqueName: \"kubernetes.io/projected/39b77803-5bb8-436f-8ca6-bc4297ea81e8-kube-api-access-frcsl\") pod \"nmstate-metrics-7f7f7578db-f64ct\" (UID: \"39b77803-5bb8-436f-8ca6-bc4297ea81e8\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-f64ct" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.484754 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/80988eb2-1173-47de-8699-b7e7bd1f63b7-nmstate-lock\") pod \"nmstate-handler-t7j8b\" (UID: \"80988eb2-1173-47de-8699-b7e7bd1f63b7\") " pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.484769 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8phd\" (UniqueName: \"kubernetes.io/projected/80988eb2-1173-47de-8699-b7e7bd1f63b7-kube-api-access-j8phd\") pod \"nmstate-handler-t7j8b\" (UID: \"80988eb2-1173-47de-8699-b7e7bd1f63b7\") " pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.484789 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/80988eb2-1173-47de-8699-b7e7bd1f63b7-dbus-socket\") pod \"nmstate-handler-t7j8b\" (UID: \"80988eb2-1173-47de-8699-b7e7bd1f63b7\") " pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.484906 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/80988eb2-1173-47de-8699-b7e7bd1f63b7-ovs-socket\") pod \"nmstate-handler-t7j8b\" (UID: \"80988eb2-1173-47de-8699-b7e7bd1f63b7\") " pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.484946 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d89bee79-f8ff-4b46-98d0-b76671baf508-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-nm6nc\" (UID: \"d89bee79-f8ff-4b46-98d0-b76671baf508\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.507740 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4c5aa8ca-24cf-4ac8-937d-349e4d15eae2-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-vtqqs\" (UID: \"4c5aa8ca-24cf-4ac8-937d-349e4d15eae2\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.516997 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc"] Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.551089 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frcsl\" (UniqueName: \"kubernetes.io/projected/39b77803-5bb8-436f-8ca6-bc4297ea81e8-kube-api-access-frcsl\") pod \"nmstate-metrics-7f7f7578db-f64ct\" (UID: \"39b77803-5bb8-436f-8ca6-bc4297ea81e8\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-f64ct" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.555043 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdlpc\" (UniqueName: \"kubernetes.io/projected/4c5aa8ca-24cf-4ac8-937d-349e4d15eae2-kube-api-access-fdlpc\") pod \"nmstate-webhook-f8fb84555-vtqqs\" (UID: \"4c5aa8ca-24cf-4ac8-937d-349e4d15eae2\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.566823 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-f64ct" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.585764 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d89bee79-f8ff-4b46-98d0-b76671baf508-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-nm6nc\" (UID: \"d89bee79-f8ff-4b46-98d0-b76671baf508\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.585985 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d89bee79-f8ff-4b46-98d0-b76671baf508-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-nm6nc\" (UID: \"d89bee79-f8ff-4b46-98d0-b76671baf508\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" Dec 11 08:50:42 crc kubenswrapper[4629]: E1211 08:50:42.586094 4629 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 11 08:50:42 crc kubenswrapper[4629]: E1211 08:50:42.586176 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d89bee79-f8ff-4b46-98d0-b76671baf508-plugin-serving-cert podName:d89bee79-f8ff-4b46-98d0-b76671baf508 nodeName:}" failed. No retries permitted until 2025-12-11 08:50:43.086146584 +0000 UTC m=+831.151562192 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/d89bee79-f8ff-4b46-98d0-b76671baf508-plugin-serving-cert") pod "nmstate-console-plugin-6ff7998486-nm6nc" (UID: "d89bee79-f8ff-4b46-98d0-b76671baf508") : secret "plugin-serving-cert" not found Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.586203 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmcr6\" (UniqueName: \"kubernetes.io/projected/d89bee79-f8ff-4b46-98d0-b76671baf508-kube-api-access-dmcr6\") pod \"nmstate-console-plugin-6ff7998486-nm6nc\" (UID: \"d89bee79-f8ff-4b46-98d0-b76671baf508\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.586269 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/80988eb2-1173-47de-8699-b7e7bd1f63b7-nmstate-lock\") pod \"nmstate-handler-t7j8b\" (UID: \"80988eb2-1173-47de-8699-b7e7bd1f63b7\") " pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.586298 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8phd\" (UniqueName: \"kubernetes.io/projected/80988eb2-1173-47de-8699-b7e7bd1f63b7-kube-api-access-j8phd\") pod \"nmstate-handler-t7j8b\" (UID: \"80988eb2-1173-47de-8699-b7e7bd1f63b7\") " pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.586327 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/80988eb2-1173-47de-8699-b7e7bd1f63b7-dbus-socket\") pod \"nmstate-handler-t7j8b\" (UID: \"80988eb2-1173-47de-8699-b7e7bd1f63b7\") " pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.586359 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/80988eb2-1173-47de-8699-b7e7bd1f63b7-ovs-socket\") pod \"nmstate-handler-t7j8b\" (UID: \"80988eb2-1173-47de-8699-b7e7bd1f63b7\") " pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.586407 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/80988eb2-1173-47de-8699-b7e7bd1f63b7-ovs-socket\") pod \"nmstate-handler-t7j8b\" (UID: \"80988eb2-1173-47de-8699-b7e7bd1f63b7\") " pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.586411 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.586649 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/80988eb2-1173-47de-8699-b7e7bd1f63b7-nmstate-lock\") pod \"nmstate-handler-t7j8b\" (UID: \"80988eb2-1173-47de-8699-b7e7bd1f63b7\") " pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.586975 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/80988eb2-1173-47de-8699-b7e7bd1f63b7-dbus-socket\") pod \"nmstate-handler-t7j8b\" (UID: \"80988eb2-1173-47de-8699-b7e7bd1f63b7\") " pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.588627 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d89bee79-f8ff-4b46-98d0-b76671baf508-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-nm6nc\" (UID: \"d89bee79-f8ff-4b46-98d0-b76671baf508\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.615195 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8phd\" (UniqueName: \"kubernetes.io/projected/80988eb2-1173-47de-8699-b7e7bd1f63b7-kube-api-access-j8phd\") pod \"nmstate-handler-t7j8b\" (UID: \"80988eb2-1173-47de-8699-b7e7bd1f63b7\") " pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.615678 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmcr6\" (UniqueName: \"kubernetes.io/projected/d89bee79-f8ff-4b46-98d0-b76671baf508-kube-api-access-dmcr6\") pod \"nmstate-console-plugin-6ff7998486-nm6nc\" (UID: \"d89bee79-f8ff-4b46-98d0-b76671baf508\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.720959 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-67c44bb664-h67h7"] Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.722001 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.765978 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-67c44bb664-h67h7"] Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.795821 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wfsm\" (UniqueName: \"kubernetes.io/projected/9c531fe9-9ea8-463a-81e4-c632e63bb809-kube-api-access-5wfsm\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.796052 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9c531fe9-9ea8-463a-81e4-c632e63bb809-console-oauth-config\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.796109 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9c531fe9-9ea8-463a-81e4-c632e63bb809-service-ca\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.796180 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9c531fe9-9ea8-463a-81e4-c632e63bb809-oauth-serving-cert\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.796205 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9c531fe9-9ea8-463a-81e4-c632e63bb809-console-serving-cert\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.796271 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c531fe9-9ea8-463a-81e4-c632e63bb809-trusted-ca-bundle\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.796296 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9c531fe9-9ea8-463a-81e4-c632e63bb809-console-config\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.898005 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c531fe9-9ea8-463a-81e4-c632e63bb809-trusted-ca-bundle\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.899654 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9c531fe9-9ea8-463a-81e4-c632e63bb809-console-config\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.899390 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c531fe9-9ea8-463a-81e4-c632e63bb809-trusted-ca-bundle\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.900485 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9c531fe9-9ea8-463a-81e4-c632e63bb809-console-config\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.901007 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wfsm\" (UniqueName: \"kubernetes.io/projected/9c531fe9-9ea8-463a-81e4-c632e63bb809-kube-api-access-5wfsm\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.901801 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9c531fe9-9ea8-463a-81e4-c632e63bb809-console-oauth-config\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.902028 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9c531fe9-9ea8-463a-81e4-c632e63bb809-service-ca\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.902201 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9c531fe9-9ea8-463a-81e4-c632e63bb809-oauth-serving-cert\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.902334 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9c531fe9-9ea8-463a-81e4-c632e63bb809-console-serving-cert\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.903792 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9c531fe9-9ea8-463a-81e4-c632e63bb809-service-ca\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.910057 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.910569 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9c531fe9-9ea8-463a-81e4-c632e63bb809-console-oauth-config\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.913870 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9c531fe9-9ea8-463a-81e4-c632e63bb809-oauth-serving-cert\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.921634 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9c531fe9-9ea8-463a-81e4-c632e63bb809-console-serving-cert\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:42 crc kubenswrapper[4629]: W1211 08:50:42.926950 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80988eb2_1173_47de_8699_b7e7bd1f63b7.slice/crio-9963dd778f125f611d00b3821c28b454abc986acbca30391352c8362ab6ca7af WatchSource:0}: Error finding container 9963dd778f125f611d00b3821c28b454abc986acbca30391352c8362ab6ca7af: Status 404 returned error can't find the container with id 9963dd778f125f611d00b3821c28b454abc986acbca30391352c8362ab6ca7af Dec 11 08:50:42 crc kubenswrapper[4629]: I1211 08:50:42.927769 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wfsm\" (UniqueName: \"kubernetes.io/projected/9c531fe9-9ea8-463a-81e4-c632e63bb809-kube-api-access-5wfsm\") pod \"console-67c44bb664-h67h7\" (UID: \"9c531fe9-9ea8-463a-81e4-c632e63bb809\") " pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:43 crc kubenswrapper[4629]: I1211 08:50:43.096798 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:43 crc kubenswrapper[4629]: I1211 08:50:43.105918 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d89bee79-f8ff-4b46-98d0-b76671baf508-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-nm6nc\" (UID: \"d89bee79-f8ff-4b46-98d0-b76671baf508\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" Dec 11 08:50:43 crc kubenswrapper[4629]: I1211 08:50:43.115180 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d89bee79-f8ff-4b46-98d0-b76671baf508-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-nm6nc\" (UID: \"d89bee79-f8ff-4b46-98d0-b76671baf508\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" Dec 11 08:50:43 crc kubenswrapper[4629]: I1211 08:50:43.115481 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" Dec 11 08:50:43 crc kubenswrapper[4629]: I1211 08:50:43.424766 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs"] Dec 11 08:50:43 crc kubenswrapper[4629]: I1211 08:50:43.685068 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-f64ct"] Dec 11 08:50:43 crc kubenswrapper[4629]: I1211 08:50:43.780275 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-67c44bb664-h67h7"] Dec 11 08:50:43 crc kubenswrapper[4629]: I1211 08:50:43.840963 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc"] Dec 11 08:50:43 crc kubenswrapper[4629]: I1211 08:50:43.843425 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-t7j8b" event={"ID":"80988eb2-1173-47de-8699-b7e7bd1f63b7","Type":"ContainerStarted","Data":"9963dd778f125f611d00b3821c28b454abc986acbca30391352c8362ab6ca7af"} Dec 11 08:50:43 crc kubenswrapper[4629]: I1211 08:50:43.845416 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-67c44bb664-h67h7" event={"ID":"9c531fe9-9ea8-463a-81e4-c632e63bb809","Type":"ContainerStarted","Data":"4bf1ebf157b99f2f04c36e35d90e7cfbe33953e445df646558f94be41bf2bf52"} Dec 11 08:50:43 crc kubenswrapper[4629]: I1211 08:50:43.848234 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" event={"ID":"4c5aa8ca-24cf-4ac8-937d-349e4d15eae2","Type":"ContainerStarted","Data":"203aa7e8fc1156f6267a2571cfd6b369f65690235eb1933fe730818f6ee795a0"} Dec 11 08:50:43 crc kubenswrapper[4629]: W1211 08:50:43.850446 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd89bee79_f8ff_4b46_98d0_b76671baf508.slice/crio-c60d3b7c4599d9d575831f2330fc5f893d4cf771966305be780e491648f884fa WatchSource:0}: Error finding container c60d3b7c4599d9d575831f2330fc5f893d4cf771966305be780e491648f884fa: Status 404 returned error can't find the container with id c60d3b7c4599d9d575831f2330fc5f893d4cf771966305be780e491648f884fa Dec 11 08:50:43 crc kubenswrapper[4629]: I1211 08:50:43.852546 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-f64ct" event={"ID":"39b77803-5bb8-436f-8ca6-bc4297ea81e8","Type":"ContainerStarted","Data":"e7a02910fc40f94494572f08414ac44ee36fa6a9d5d5aceba4db59d50975fa5e"} Dec 11 08:50:44 crc kubenswrapper[4629]: I1211 08:50:44.829934 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:44 crc kubenswrapper[4629]: I1211 08:50:44.861352 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" event={"ID":"d89bee79-f8ff-4b46-98d0-b76671baf508","Type":"ContainerStarted","Data":"c60d3b7c4599d9d575831f2330fc5f893d4cf771966305be780e491648f884fa"} Dec 11 08:50:44 crc kubenswrapper[4629]: I1211 08:50:44.871387 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:45 crc kubenswrapper[4629]: I1211 08:50:45.057361 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z2vwh"] Dec 11 08:50:45 crc kubenswrapper[4629]: I1211 08:50:45.866260 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z2vwh" podUID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" containerName="registry-server" containerID="cri-o://eafa7c9efe588b9ed4df2e19d4ccae2731febcb9ef5c3e9daa96625fc1589cc4" gracePeriod=2 Dec 11 08:50:46 crc kubenswrapper[4629]: I1211 08:50:46.877236 4629 generic.go:334] "Generic (PLEG): container finished" podID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" containerID="eafa7c9efe588b9ed4df2e19d4ccae2731febcb9ef5c3e9daa96625fc1589cc4" exitCode=0 Dec 11 08:50:46 crc kubenswrapper[4629]: I1211 08:50:46.877300 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2vwh" event={"ID":"5fe1b4ed-87d1-4802-8350-59e3ed522eeb","Type":"ContainerDied","Data":"eafa7c9efe588b9ed4df2e19d4ccae2731febcb9ef5c3e9daa96625fc1589cc4"} Dec 11 08:50:46 crc kubenswrapper[4629]: I1211 08:50:46.880471 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-67c44bb664-h67h7" event={"ID":"9c531fe9-9ea8-463a-81e4-c632e63bb809","Type":"ContainerStarted","Data":"9cbda56c254a4eb7bacaace022d125b57bed1dd48085dd72fc700166e75d9e1a"} Dec 11 08:50:46 crc kubenswrapper[4629]: I1211 08:50:46.906357 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-67c44bb664-h67h7" podStartSLOduration=4.9063261350000005 podStartE2EDuration="4.906326135s" podCreationTimestamp="2025-12-11 08:50:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:50:46.900401278 +0000 UTC m=+834.965816906" watchObservedRunningTime="2025-12-11 08:50:46.906326135 +0000 UTC m=+834.971741753" Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.331356 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.376103 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md2l4\" (UniqueName: \"kubernetes.io/projected/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-kube-api-access-md2l4\") pod \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\" (UID: \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\") " Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.376167 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-utilities\") pod \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\" (UID: \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\") " Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.376216 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-catalog-content\") pod \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\" (UID: \"5fe1b4ed-87d1-4802-8350-59e3ed522eeb\") " Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.377637 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-utilities" (OuterVolumeSpecName: "utilities") pod "5fe1b4ed-87d1-4802-8350-59e3ed522eeb" (UID: "5fe1b4ed-87d1-4802-8350-59e3ed522eeb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.389657 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-kube-api-access-md2l4" (OuterVolumeSpecName: "kube-api-access-md2l4") pod "5fe1b4ed-87d1-4802-8350-59e3ed522eeb" (UID: "5fe1b4ed-87d1-4802-8350-59e3ed522eeb"). InnerVolumeSpecName "kube-api-access-md2l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.477549 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md2l4\" (UniqueName: \"kubernetes.io/projected/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-kube-api-access-md2l4\") on node \"crc\" DevicePath \"\"" Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.477583 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.522263 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fe1b4ed-87d1-4802-8350-59e3ed522eeb" (UID: "5fe1b4ed-87d1-4802-8350-59e3ed522eeb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.579364 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe1b4ed-87d1-4802-8350-59e3ed522eeb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.887784 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2vwh" event={"ID":"5fe1b4ed-87d1-4802-8350-59e3ed522eeb","Type":"ContainerDied","Data":"c54a69aafca0ff7cac1e480daf7166d752311d8c3aef05ea03bd0aecfeedab08"} Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.888174 4629 scope.go:117] "RemoveContainer" containerID="eafa7c9efe588b9ed4df2e19d4ccae2731febcb9ef5c3e9daa96625fc1589cc4" Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.887820 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2vwh" Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.922931 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z2vwh"] Dec 11 08:50:47 crc kubenswrapper[4629]: I1211 08:50:47.929135 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z2vwh"] Dec 11 08:50:48 crc kubenswrapper[4629]: I1211 08:50:48.207286 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" path="/var/lib/kubelet/pods/5fe1b4ed-87d1-4802-8350-59e3ed522eeb/volumes" Dec 11 08:50:49 crc kubenswrapper[4629]: I1211 08:50:49.394737 4629 scope.go:117] "RemoveContainer" containerID="393487a3acae4fed319c25a331cce5684717d880714a684a15b5c27c42ac7c66" Dec 11 08:50:49 crc kubenswrapper[4629]: I1211 08:50:49.460455 4629 scope.go:117] "RemoveContainer" containerID="4e9d2c90cf7f173d71fe44f80b385e666ad5bc39d1f02a984263f9566b151aec" Dec 11 08:50:50 crc kubenswrapper[4629]: I1211 08:50:50.907957 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-t7j8b" event={"ID":"80988eb2-1173-47de-8699-b7e7bd1f63b7","Type":"ContainerStarted","Data":"d3609b749e688973feddd1e1d26b0649aaae8754a0af353bb2d9a2adbe93c59a"} Dec 11 08:50:50 crc kubenswrapper[4629]: I1211 08:50:50.910265 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:50:50 crc kubenswrapper[4629]: I1211 08:50:50.915755 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" event={"ID":"d89bee79-f8ff-4b46-98d0-b76671baf508","Type":"ContainerStarted","Data":"aad581e13333effc1943d68e9237d1cbda2a22a3ff9e3ab0e92619cb80a71453"} Dec 11 08:50:50 crc kubenswrapper[4629]: I1211 08:50:50.918650 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" event={"ID":"4c5aa8ca-24cf-4ac8-937d-349e4d15eae2","Type":"ContainerStarted","Data":"00f377dd4223ce93be98ea40af251a222471c01936fd3ae10a79ad7db3fc06e4"} Dec 11 08:50:50 crc kubenswrapper[4629]: I1211 08:50:50.918887 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" Dec 11 08:50:50 crc kubenswrapper[4629]: I1211 08:50:50.920734 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-f64ct" event={"ID":"39b77803-5bb8-436f-8ca6-bc4297ea81e8","Type":"ContainerStarted","Data":"891880ebd22755f893796c9da9dc8f00756351dc4529fc14245504294fa6cb27"} Dec 11 08:50:50 crc kubenswrapper[4629]: I1211 08:50:50.929326 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-t7j8b" podStartSLOduration=2.254941671 podStartE2EDuration="8.929310104s" podCreationTimestamp="2025-12-11 08:50:42 +0000 UTC" firstStartedPulling="2025-12-11 08:50:42.928329084 +0000 UTC m=+830.993744692" lastFinishedPulling="2025-12-11 08:50:49.602697517 +0000 UTC m=+837.668113125" observedRunningTime="2025-12-11 08:50:50.926754512 +0000 UTC m=+838.992170120" watchObservedRunningTime="2025-12-11 08:50:50.929310104 +0000 UTC m=+838.994725712" Dec 11 08:50:50 crc kubenswrapper[4629]: I1211 08:50:50.945365 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-nm6nc" podStartSLOduration=3.401389469 podStartE2EDuration="8.94534198s" podCreationTimestamp="2025-12-11 08:50:42 +0000 UTC" firstStartedPulling="2025-12-11 08:50:43.853387935 +0000 UTC m=+831.918803543" lastFinishedPulling="2025-12-11 08:50:49.397340446 +0000 UTC m=+837.462756054" observedRunningTime="2025-12-11 08:50:50.943728278 +0000 UTC m=+839.009143886" watchObservedRunningTime="2025-12-11 08:50:50.94534198 +0000 UTC m=+839.010757588" Dec 11 08:50:52 crc kubenswrapper[4629]: I1211 08:50:52.222023 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" podStartSLOduration=4.042206647 podStartE2EDuration="10.222001899s" podCreationTimestamp="2025-12-11 08:50:42 +0000 UTC" firstStartedPulling="2025-12-11 08:50:43.463705044 +0000 UTC m=+831.529120652" lastFinishedPulling="2025-12-11 08:50:49.643500296 +0000 UTC m=+837.708915904" observedRunningTime="2025-12-11 08:50:50.968780109 +0000 UTC m=+839.034195717" watchObservedRunningTime="2025-12-11 08:50:52.222001899 +0000 UTC m=+840.287417507" Dec 11 08:50:52 crc kubenswrapper[4629]: I1211 08:50:52.932131 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-f64ct" event={"ID":"39b77803-5bb8-436f-8ca6-bc4297ea81e8","Type":"ContainerStarted","Data":"ff0518d6460737177f21807ab9bb1ffbd02f048b8cec21a0d24d17429c0bad1a"} Dec 11 08:50:52 crc kubenswrapper[4629]: I1211 08:50:52.953747 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-f64ct" podStartSLOduration=2.218592113 podStartE2EDuration="10.953720285s" podCreationTimestamp="2025-12-11 08:50:42 +0000 UTC" firstStartedPulling="2025-12-11 08:50:43.714547693 +0000 UTC m=+831.779963301" lastFinishedPulling="2025-12-11 08:50:52.449675865 +0000 UTC m=+840.515091473" observedRunningTime="2025-12-11 08:50:52.951571578 +0000 UTC m=+841.016987206" watchObservedRunningTime="2025-12-11 08:50:52.953720285 +0000 UTC m=+841.019135893" Dec 11 08:50:53 crc kubenswrapper[4629]: I1211 08:50:53.099018 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:53 crc kubenswrapper[4629]: I1211 08:50:53.099056 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:53 crc kubenswrapper[4629]: I1211 08:50:53.103456 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:53 crc kubenswrapper[4629]: I1211 08:50:53.940527 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-67c44bb664-h67h7" Dec 11 08:50:53 crc kubenswrapper[4629]: I1211 08:50:53.994879 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-lhszp"] Dec 11 08:50:57 crc kubenswrapper[4629]: I1211 08:50:57.932995 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-t7j8b" Dec 11 08:51:02 crc kubenswrapper[4629]: I1211 08:51:02.594339 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-f8fb84555-vtqqs" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.464099 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5jlbd"] Dec 11 08:51:11 crc kubenswrapper[4629]: E1211 08:51:11.464935 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" containerName="extract-content" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.464948 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" containerName="extract-content" Dec 11 08:51:11 crc kubenswrapper[4629]: E1211 08:51:11.464962 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" containerName="registry-server" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.464968 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" containerName="registry-server" Dec 11 08:51:11 crc kubenswrapper[4629]: E1211 08:51:11.464983 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" containerName="extract-utilities" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.464989 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" containerName="extract-utilities" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.465102 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe1b4ed-87d1-4802-8350-59e3ed522eeb" containerName="registry-server" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.465889 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.466157 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5jlbd"] Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.470293 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62bfaf65-a734-4a07-960c-6a3363325479-utilities\") pod \"redhat-marketplace-5jlbd\" (UID: \"62bfaf65-a734-4a07-960c-6a3363325479\") " pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.470340 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62bfaf65-a734-4a07-960c-6a3363325479-catalog-content\") pod \"redhat-marketplace-5jlbd\" (UID: \"62bfaf65-a734-4a07-960c-6a3363325479\") " pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.470403 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdbxn\" (UniqueName: \"kubernetes.io/projected/62bfaf65-a734-4a07-960c-6a3363325479-kube-api-access-xdbxn\") pod \"redhat-marketplace-5jlbd\" (UID: \"62bfaf65-a734-4a07-960c-6a3363325479\") " pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.571195 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdbxn\" (UniqueName: \"kubernetes.io/projected/62bfaf65-a734-4a07-960c-6a3363325479-kube-api-access-xdbxn\") pod \"redhat-marketplace-5jlbd\" (UID: \"62bfaf65-a734-4a07-960c-6a3363325479\") " pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.571520 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62bfaf65-a734-4a07-960c-6a3363325479-utilities\") pod \"redhat-marketplace-5jlbd\" (UID: \"62bfaf65-a734-4a07-960c-6a3363325479\") " pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.571538 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62bfaf65-a734-4a07-960c-6a3363325479-catalog-content\") pod \"redhat-marketplace-5jlbd\" (UID: \"62bfaf65-a734-4a07-960c-6a3363325479\") " pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.572049 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62bfaf65-a734-4a07-960c-6a3363325479-catalog-content\") pod \"redhat-marketplace-5jlbd\" (UID: \"62bfaf65-a734-4a07-960c-6a3363325479\") " pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.572538 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62bfaf65-a734-4a07-960c-6a3363325479-utilities\") pod \"redhat-marketplace-5jlbd\" (UID: \"62bfaf65-a734-4a07-960c-6a3363325479\") " pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.592342 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdbxn\" (UniqueName: \"kubernetes.io/projected/62bfaf65-a734-4a07-960c-6a3363325479-kube-api-access-xdbxn\") pod \"redhat-marketplace-5jlbd\" (UID: \"62bfaf65-a734-4a07-960c-6a3363325479\") " pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:11 crc kubenswrapper[4629]: I1211 08:51:11.837969 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:12 crc kubenswrapper[4629]: I1211 08:51:12.315632 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5jlbd"] Dec 11 08:51:12 crc kubenswrapper[4629]: W1211 08:51:12.318833 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62bfaf65_a734_4a07_960c_6a3363325479.slice/crio-95b001d06fffb8afb768a070206cab8d07edcfa8d0c8fb3a388ea83d5951481f WatchSource:0}: Error finding container 95b001d06fffb8afb768a070206cab8d07edcfa8d0c8fb3a388ea83d5951481f: Status 404 returned error can't find the container with id 95b001d06fffb8afb768a070206cab8d07edcfa8d0c8fb3a388ea83d5951481f Dec 11 08:51:13 crc kubenswrapper[4629]: I1211 08:51:13.040411 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5jlbd" event={"ID":"62bfaf65-a734-4a07-960c-6a3363325479","Type":"ContainerStarted","Data":"95b001d06fffb8afb768a070206cab8d07edcfa8d0c8fb3a388ea83d5951481f"} Dec 11 08:51:14 crc kubenswrapper[4629]: I1211 08:51:14.046045 4629 generic.go:334] "Generic (PLEG): container finished" podID="62bfaf65-a734-4a07-960c-6a3363325479" containerID="318b69ca652d56cb6e8daefa66b5dcefec827375dc25a1e3b2f777134d5bc2cf" exitCode=0 Dec 11 08:51:14 crc kubenswrapper[4629]: I1211 08:51:14.046123 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5jlbd" event={"ID":"62bfaf65-a734-4a07-960c-6a3363325479","Type":"ContainerDied","Data":"318b69ca652d56cb6e8daefa66b5dcefec827375dc25a1e3b2f777134d5bc2cf"} Dec 11 08:51:15 crc kubenswrapper[4629]: I1211 08:51:15.054351 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5jlbd" event={"ID":"62bfaf65-a734-4a07-960c-6a3363325479","Type":"ContainerStarted","Data":"902c3a04b8391663cb2f1ec7c2d918ad717a8f24f033644b5a32faa10ce2915f"} Dec 11 08:51:16 crc kubenswrapper[4629]: I1211 08:51:16.076456 4629 generic.go:334] "Generic (PLEG): container finished" podID="62bfaf65-a734-4a07-960c-6a3363325479" containerID="902c3a04b8391663cb2f1ec7c2d918ad717a8f24f033644b5a32faa10ce2915f" exitCode=0 Dec 11 08:51:16 crc kubenswrapper[4629]: I1211 08:51:16.076520 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5jlbd" event={"ID":"62bfaf65-a734-4a07-960c-6a3363325479","Type":"ContainerDied","Data":"902c3a04b8391663cb2f1ec7c2d918ad717a8f24f033644b5a32faa10ce2915f"} Dec 11 08:51:16 crc kubenswrapper[4629]: I1211 08:51:16.881973 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs"] Dec 11 08:51:16 crc kubenswrapper[4629]: I1211 08:51:16.883585 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:16 crc kubenswrapper[4629]: I1211 08:51:16.886328 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 11 08:51:16 crc kubenswrapper[4629]: I1211 08:51:16.893784 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs"] Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.038323 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4k9g\" (UniqueName: \"kubernetes.io/projected/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-kube-api-access-w4k9g\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs\" (UID: \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.038410 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs\" (UID: \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.038473 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs\" (UID: \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.084288 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5jlbd" event={"ID":"62bfaf65-a734-4a07-960c-6a3363325479","Type":"ContainerStarted","Data":"d44c500ccfb81eb5c0e2da39dc50c57545fce171a084d7881b5428135400c5a0"} Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.101942 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5jlbd" podStartSLOduration=3.611671449 podStartE2EDuration="6.101914827s" podCreationTimestamp="2025-12-11 08:51:11 +0000 UTC" firstStartedPulling="2025-12-11 08:51:14.048871434 +0000 UTC m=+862.114287042" lastFinishedPulling="2025-12-11 08:51:16.539114812 +0000 UTC m=+864.604530420" observedRunningTime="2025-12-11 08:51:17.099814861 +0000 UTC m=+865.165230479" watchObservedRunningTime="2025-12-11 08:51:17.101914827 +0000 UTC m=+865.167330445" Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.139098 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4k9g\" (UniqueName: \"kubernetes.io/projected/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-kube-api-access-w4k9g\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs\" (UID: \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.139165 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs\" (UID: \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.139192 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs\" (UID: \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.139620 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs\" (UID: \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.139994 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs\" (UID: \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.164991 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4k9g\" (UniqueName: \"kubernetes.io/projected/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-kube-api-access-w4k9g\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs\" (UID: \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.199159 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:17 crc kubenswrapper[4629]: I1211 08:51:17.395861 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs"] Dec 11 08:51:18 crc kubenswrapper[4629]: I1211 08:51:18.092552 4629 generic.go:334] "Generic (PLEG): container finished" podID="dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" containerID="19c272e3d8366f99b74adcd6ee3c9394e259f16bf4bbdfdc789d566a22ffafef" exitCode=0 Dec 11 08:51:18 crc kubenswrapper[4629]: I1211 08:51:18.092647 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" event={"ID":"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377","Type":"ContainerDied","Data":"19c272e3d8366f99b74adcd6ee3c9394e259f16bf4bbdfdc789d566a22ffafef"} Dec 11 08:51:18 crc kubenswrapper[4629]: I1211 08:51:18.092687 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" event={"ID":"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377","Type":"ContainerStarted","Data":"83bcb96784790c9b479a160c3a4a6a8045a8936d864657b337a3377951086902"} Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.030705 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-lhszp" podUID="9d819e7a-6f7a-4f6e-bbba-4b334546f172" containerName="console" containerID="cri-o://751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3" gracePeriod=15 Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.460556 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-lhszp_9d819e7a-6f7a-4f6e-bbba-4b334546f172/console/0.log" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.460858 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.566660 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-oauth-serving-cert\") pod \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.566803 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-trusted-ca-bundle\") pod \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.566841 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-serving-cert\") pod \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.566930 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-oauth-config\") pod \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.566982 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-config\") pod \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.567014 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2kbq\" (UniqueName: \"kubernetes.io/projected/9d819e7a-6f7a-4f6e-bbba-4b334546f172-kube-api-access-f2kbq\") pod \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.567038 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-service-ca\") pod \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\" (UID: \"9d819e7a-6f7a-4f6e-bbba-4b334546f172\") " Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.567923 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9d819e7a-6f7a-4f6e-bbba-4b334546f172" (UID: "9d819e7a-6f7a-4f6e-bbba-4b334546f172"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.567931 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-config" (OuterVolumeSpecName: "console-config") pod "9d819e7a-6f7a-4f6e-bbba-4b334546f172" (UID: "9d819e7a-6f7a-4f6e-bbba-4b334546f172"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.568250 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-service-ca" (OuterVolumeSpecName: "service-ca") pod "9d819e7a-6f7a-4f6e-bbba-4b334546f172" (UID: "9d819e7a-6f7a-4f6e-bbba-4b334546f172"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.568508 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9d819e7a-6f7a-4f6e-bbba-4b334546f172" (UID: "9d819e7a-6f7a-4f6e-bbba-4b334546f172"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.569897 4629 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.570004 4629 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.570072 4629 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.570146 4629 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d819e7a-6f7a-4f6e-bbba-4b334546f172-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.578175 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9d819e7a-6f7a-4f6e-bbba-4b334546f172" (UID: "9d819e7a-6f7a-4f6e-bbba-4b334546f172"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.579025 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9d819e7a-6f7a-4f6e-bbba-4b334546f172" (UID: "9d819e7a-6f7a-4f6e-bbba-4b334546f172"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.579767 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d819e7a-6f7a-4f6e-bbba-4b334546f172-kube-api-access-f2kbq" (OuterVolumeSpecName: "kube-api-access-f2kbq") pod "9d819e7a-6f7a-4f6e-bbba-4b334546f172" (UID: "9d819e7a-6f7a-4f6e-bbba-4b334546f172"). InnerVolumeSpecName "kube-api-access-f2kbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.671766 4629 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.671808 4629 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d819e7a-6f7a-4f6e-bbba-4b334546f172-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:19 crc kubenswrapper[4629]: I1211 08:51:19.671824 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2kbq\" (UniqueName: \"kubernetes.io/projected/9d819e7a-6f7a-4f6e-bbba-4b334546f172-kube-api-access-f2kbq\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:20 crc kubenswrapper[4629]: I1211 08:51:20.785818 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-lhszp_9d819e7a-6f7a-4f6e-bbba-4b334546f172/console/0.log" Dec 11 08:51:20 crc kubenswrapper[4629]: I1211 08:51:20.785882 4629 generic.go:334] "Generic (PLEG): container finished" podID="9d819e7a-6f7a-4f6e-bbba-4b334546f172" containerID="751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3" exitCode=2 Dec 11 08:51:20 crc kubenswrapper[4629]: I1211 08:51:20.787641 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-lhszp" event={"ID":"9d819e7a-6f7a-4f6e-bbba-4b334546f172","Type":"ContainerDied","Data":"751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3"} Dec 11 08:51:20 crc kubenswrapper[4629]: I1211 08:51:20.787698 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-lhszp" event={"ID":"9d819e7a-6f7a-4f6e-bbba-4b334546f172","Type":"ContainerDied","Data":"ea5b81c4e7c45ef5d714258359fd60e70ca8d5febeb8e08f62cff7dfd65067ec"} Dec 11 08:51:20 crc kubenswrapper[4629]: I1211 08:51:20.787720 4629 scope.go:117] "RemoveContainer" containerID="751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3" Dec 11 08:51:20 crc kubenswrapper[4629]: I1211 08:51:20.787901 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-lhszp" Dec 11 08:51:20 crc kubenswrapper[4629]: I1211 08:51:20.831398 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-lhszp"] Dec 11 08:51:20 crc kubenswrapper[4629]: I1211 08:51:20.835424 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-lhszp"] Dec 11 08:51:21 crc kubenswrapper[4629]: I1211 08:51:21.283580 4629 scope.go:117] "RemoveContainer" containerID="751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3" Dec 11 08:51:21 crc kubenswrapper[4629]: E1211 08:51:21.284231 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3\": container with ID starting with 751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3 not found: ID does not exist" containerID="751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3" Dec 11 08:51:21 crc kubenswrapper[4629]: I1211 08:51:21.284267 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3"} err="failed to get container status \"751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3\": rpc error: code = NotFound desc = could not find container \"751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3\": container with ID starting with 751e239be77a656e55384121ce97d22fa7370f77a7834d4f2bd936fbf0bddbe3 not found: ID does not exist" Dec 11 08:51:21 crc kubenswrapper[4629]: I1211 08:51:21.795500 4629 generic.go:334] "Generic (PLEG): container finished" podID="dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" containerID="9e73a8c138ec6b357bf3652b0e53e8b39e9abd96c63abe1064cb19df7e919b30" exitCode=0 Dec 11 08:51:21 crc kubenswrapper[4629]: I1211 08:51:21.795534 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" event={"ID":"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377","Type":"ContainerDied","Data":"9e73a8c138ec6b357bf3652b0e53e8b39e9abd96c63abe1064cb19df7e919b30"} Dec 11 08:51:21 crc kubenswrapper[4629]: I1211 08:51:21.838840 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:21 crc kubenswrapper[4629]: I1211 08:51:21.838971 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:21 crc kubenswrapper[4629]: I1211 08:51:21.893127 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:22 crc kubenswrapper[4629]: I1211 08:51:22.204685 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d819e7a-6f7a-4f6e-bbba-4b334546f172" path="/var/lib/kubelet/pods/9d819e7a-6f7a-4f6e-bbba-4b334546f172/volumes" Dec 11 08:51:22 crc kubenswrapper[4629]: I1211 08:51:22.804289 4629 generic.go:334] "Generic (PLEG): container finished" podID="dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" containerID="a57ecdd1b4a5768e8940aa6f47bfa6567ac80d89aa41693195ece20eefe60dad" exitCode=0 Dec 11 08:51:22 crc kubenswrapper[4629]: I1211 08:51:22.805839 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" event={"ID":"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377","Type":"ContainerDied","Data":"a57ecdd1b4a5768e8940aa6f47bfa6567ac80d89aa41693195ece20eefe60dad"} Dec 11 08:51:22 crc kubenswrapper[4629]: I1211 08:51:22.848940 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.122126 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.253451 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-util\") pod \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\" (UID: \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\") " Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.253542 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4k9g\" (UniqueName: \"kubernetes.io/projected/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-kube-api-access-w4k9g\") pod \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\" (UID: \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\") " Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.253634 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-bundle\") pod \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\" (UID: \"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377\") " Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.254810 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-bundle" (OuterVolumeSpecName: "bundle") pod "dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" (UID: "dd7dc11b-1ceb-45c3-bd58-ef07d3da9377"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.266496 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-kube-api-access-w4k9g" (OuterVolumeSpecName: "kube-api-access-w4k9g") pod "dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" (UID: "dd7dc11b-1ceb-45c3-bd58-ef07d3da9377"). InnerVolumeSpecName "kube-api-access-w4k9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.266994 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-util" (OuterVolumeSpecName: "util") pod "dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" (UID: "dd7dc11b-1ceb-45c3-bd58-ef07d3da9377"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.355141 4629 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.355185 4629 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-util\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.355202 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4k9g\" (UniqueName: \"kubernetes.io/projected/dd7dc11b-1ceb-45c3-bd58-ef07d3da9377-kube-api-access-w4k9g\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.817833 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" event={"ID":"dd7dc11b-1ceb-45c3-bd58-ef07d3da9377","Type":"ContainerDied","Data":"83bcb96784790c9b479a160c3a4a6a8045a8936d864657b337a3377951086902"} Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.817926 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83bcb96784790c9b479a160c3a4a6a8045a8936d864657b337a3377951086902" Dec 11 08:51:24 crc kubenswrapper[4629]: I1211 08:51:24.818067 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs" Dec 11 08:51:25 crc kubenswrapper[4629]: I1211 08:51:25.228508 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5jlbd"] Dec 11 08:51:25 crc kubenswrapper[4629]: I1211 08:51:25.229378 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5jlbd" podUID="62bfaf65-a734-4a07-960c-6a3363325479" containerName="registry-server" containerID="cri-o://d44c500ccfb81eb5c0e2da39dc50c57545fce171a084d7881b5428135400c5a0" gracePeriod=2 Dec 11 08:51:25 crc kubenswrapper[4629]: I1211 08:51:25.825218 4629 generic.go:334] "Generic (PLEG): container finished" podID="62bfaf65-a734-4a07-960c-6a3363325479" containerID="d44c500ccfb81eb5c0e2da39dc50c57545fce171a084d7881b5428135400c5a0" exitCode=0 Dec 11 08:51:25 crc kubenswrapper[4629]: I1211 08:51:25.825246 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5jlbd" event={"ID":"62bfaf65-a734-4a07-960c-6a3363325479","Type":"ContainerDied","Data":"d44c500ccfb81eb5c0e2da39dc50c57545fce171a084d7881b5428135400c5a0"} Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.064775 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.176976 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62bfaf65-a734-4a07-960c-6a3363325479-catalog-content\") pod \"62bfaf65-a734-4a07-960c-6a3363325479\" (UID: \"62bfaf65-a734-4a07-960c-6a3363325479\") " Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.177031 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdbxn\" (UniqueName: \"kubernetes.io/projected/62bfaf65-a734-4a07-960c-6a3363325479-kube-api-access-xdbxn\") pod \"62bfaf65-a734-4a07-960c-6a3363325479\" (UID: \"62bfaf65-a734-4a07-960c-6a3363325479\") " Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.177104 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62bfaf65-a734-4a07-960c-6a3363325479-utilities\") pod \"62bfaf65-a734-4a07-960c-6a3363325479\" (UID: \"62bfaf65-a734-4a07-960c-6a3363325479\") " Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.177809 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62bfaf65-a734-4a07-960c-6a3363325479-utilities" (OuterVolumeSpecName: "utilities") pod "62bfaf65-a734-4a07-960c-6a3363325479" (UID: "62bfaf65-a734-4a07-960c-6a3363325479"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.184988 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62bfaf65-a734-4a07-960c-6a3363325479-kube-api-access-xdbxn" (OuterVolumeSpecName: "kube-api-access-xdbxn") pod "62bfaf65-a734-4a07-960c-6a3363325479" (UID: "62bfaf65-a734-4a07-960c-6a3363325479"). InnerVolumeSpecName "kube-api-access-xdbxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.240976 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62bfaf65-a734-4a07-960c-6a3363325479-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62bfaf65-a734-4a07-960c-6a3363325479" (UID: "62bfaf65-a734-4a07-960c-6a3363325479"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.279346 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62bfaf65-a734-4a07-960c-6a3363325479-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.279420 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62bfaf65-a734-4a07-960c-6a3363325479-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.279435 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdbxn\" (UniqueName: \"kubernetes.io/projected/62bfaf65-a734-4a07-960c-6a3363325479-kube-api-access-xdbxn\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.833570 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5jlbd" event={"ID":"62bfaf65-a734-4a07-960c-6a3363325479","Type":"ContainerDied","Data":"95b001d06fffb8afb768a070206cab8d07edcfa8d0c8fb3a388ea83d5951481f"} Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.833621 4629 scope.go:117] "RemoveContainer" containerID="d44c500ccfb81eb5c0e2da39dc50c57545fce171a084d7881b5428135400c5a0" Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.833733 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5jlbd" Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.860727 4629 scope.go:117] "RemoveContainer" containerID="902c3a04b8391663cb2f1ec7c2d918ad717a8f24f033644b5a32faa10ce2915f" Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.871348 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5jlbd"] Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.884241 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5jlbd"] Dec 11 08:51:26 crc kubenswrapper[4629]: I1211 08:51:26.884419 4629 scope.go:117] "RemoveContainer" containerID="318b69ca652d56cb6e8daefa66b5dcefec827375dc25a1e3b2f777134d5bc2cf" Dec 11 08:51:28 crc kubenswrapper[4629]: I1211 08:51:28.213990 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62bfaf65-a734-4a07-960c-6a3363325479" path="/var/lib/kubelet/pods/62bfaf65-a734-4a07-960c-6a3363325479/volumes" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.391269 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj"] Dec 11 08:51:34 crc kubenswrapper[4629]: E1211 08:51:34.391887 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" containerName="util" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.391903 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" containerName="util" Dec 11 08:51:34 crc kubenswrapper[4629]: E1211 08:51:34.391920 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62bfaf65-a734-4a07-960c-6a3363325479" containerName="registry-server" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.391927 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="62bfaf65-a734-4a07-960c-6a3363325479" containerName="registry-server" Dec 11 08:51:34 crc kubenswrapper[4629]: E1211 08:51:34.391940 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" containerName="extract" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.391946 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" containerName="extract" Dec 11 08:51:34 crc kubenswrapper[4629]: E1211 08:51:34.391959 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62bfaf65-a734-4a07-960c-6a3363325479" containerName="extract-utilities" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.391968 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="62bfaf65-a734-4a07-960c-6a3363325479" containerName="extract-utilities" Dec 11 08:51:34 crc kubenswrapper[4629]: E1211 08:51:34.391976 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" containerName="pull" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.391982 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" containerName="pull" Dec 11 08:51:34 crc kubenswrapper[4629]: E1211 08:51:34.391989 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d819e7a-6f7a-4f6e-bbba-4b334546f172" containerName="console" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.391997 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d819e7a-6f7a-4f6e-bbba-4b334546f172" containerName="console" Dec 11 08:51:34 crc kubenswrapper[4629]: E1211 08:51:34.392010 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62bfaf65-a734-4a07-960c-6a3363325479" containerName="extract-content" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.392016 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="62bfaf65-a734-4a07-960c-6a3363325479" containerName="extract-content" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.392151 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d819e7a-6f7a-4f6e-bbba-4b334546f172" containerName="console" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.392164 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="62bfaf65-a734-4a07-960c-6a3363325479" containerName="registry-server" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.392176 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd7dc11b-1ceb-45c3-bd58-ef07d3da9377" containerName="extract" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.392649 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.394395 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.394409 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.395637 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-rw6p7" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.396380 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.397674 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.457393 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj"] Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.492285 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b51a6d95-cf61-40c5-9d03-92c5c91c16ab-webhook-cert\") pod \"metallb-operator-controller-manager-6f5496ff8f-p5ldj\" (UID: \"b51a6d95-cf61-40c5-9d03-92c5c91c16ab\") " pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.492327 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b51a6d95-cf61-40c5-9d03-92c5c91c16ab-apiservice-cert\") pod \"metallb-operator-controller-manager-6f5496ff8f-p5ldj\" (UID: \"b51a6d95-cf61-40c5-9d03-92c5c91c16ab\") " pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.492357 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2t78\" (UniqueName: \"kubernetes.io/projected/b51a6d95-cf61-40c5-9d03-92c5c91c16ab-kube-api-access-s2t78\") pod \"metallb-operator-controller-manager-6f5496ff8f-p5ldj\" (UID: \"b51a6d95-cf61-40c5-9d03-92c5c91c16ab\") " pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.593479 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b51a6d95-cf61-40c5-9d03-92c5c91c16ab-webhook-cert\") pod \"metallb-operator-controller-manager-6f5496ff8f-p5ldj\" (UID: \"b51a6d95-cf61-40c5-9d03-92c5c91c16ab\") " pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.594614 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b51a6d95-cf61-40c5-9d03-92c5c91c16ab-apiservice-cert\") pod \"metallb-operator-controller-manager-6f5496ff8f-p5ldj\" (UID: \"b51a6d95-cf61-40c5-9d03-92c5c91c16ab\") " pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.594748 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2t78\" (UniqueName: \"kubernetes.io/projected/b51a6d95-cf61-40c5-9d03-92c5c91c16ab-kube-api-access-s2t78\") pod \"metallb-operator-controller-manager-6f5496ff8f-p5ldj\" (UID: \"b51a6d95-cf61-40c5-9d03-92c5c91c16ab\") " pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.600518 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b51a6d95-cf61-40c5-9d03-92c5c91c16ab-webhook-cert\") pod \"metallb-operator-controller-manager-6f5496ff8f-p5ldj\" (UID: \"b51a6d95-cf61-40c5-9d03-92c5c91c16ab\") " pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.603205 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b51a6d95-cf61-40c5-9d03-92c5c91c16ab-apiservice-cert\") pod \"metallb-operator-controller-manager-6f5496ff8f-p5ldj\" (UID: \"b51a6d95-cf61-40c5-9d03-92c5c91c16ab\") " pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.697579 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2t78\" (UniqueName: \"kubernetes.io/projected/b51a6d95-cf61-40c5-9d03-92c5c91c16ab-kube-api-access-s2t78\") pod \"metallb-operator-controller-manager-6f5496ff8f-p5ldj\" (UID: \"b51a6d95-cf61-40c5-9d03-92c5c91c16ab\") " pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.711601 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.794503 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-649886db6b-999qk"] Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.795666 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.798711 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.798945 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.799097 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-rr8qv" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.820456 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-649886db6b-999qk"] Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.899541 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htbtr\" (UniqueName: \"kubernetes.io/projected/210c22a1-a398-48f1-891f-21fecd9f53d5-kube-api-access-htbtr\") pod \"metallb-operator-webhook-server-649886db6b-999qk\" (UID: \"210c22a1-a398-48f1-891f-21fecd9f53d5\") " pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.899601 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/210c22a1-a398-48f1-891f-21fecd9f53d5-webhook-cert\") pod \"metallb-operator-webhook-server-649886db6b-999qk\" (UID: \"210c22a1-a398-48f1-891f-21fecd9f53d5\") " pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:51:34 crc kubenswrapper[4629]: I1211 08:51:34.899634 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/210c22a1-a398-48f1-891f-21fecd9f53d5-apiservice-cert\") pod \"metallb-operator-webhook-server-649886db6b-999qk\" (UID: \"210c22a1-a398-48f1-891f-21fecd9f53d5\") " pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:51:35 crc kubenswrapper[4629]: I1211 08:51:35.002283 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htbtr\" (UniqueName: \"kubernetes.io/projected/210c22a1-a398-48f1-891f-21fecd9f53d5-kube-api-access-htbtr\") pod \"metallb-operator-webhook-server-649886db6b-999qk\" (UID: \"210c22a1-a398-48f1-891f-21fecd9f53d5\") " pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:51:35 crc kubenswrapper[4629]: I1211 08:51:35.002670 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/210c22a1-a398-48f1-891f-21fecd9f53d5-webhook-cert\") pod \"metallb-operator-webhook-server-649886db6b-999qk\" (UID: \"210c22a1-a398-48f1-891f-21fecd9f53d5\") " pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:51:35 crc kubenswrapper[4629]: I1211 08:51:35.002710 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/210c22a1-a398-48f1-891f-21fecd9f53d5-apiservice-cert\") pod \"metallb-operator-webhook-server-649886db6b-999qk\" (UID: \"210c22a1-a398-48f1-891f-21fecd9f53d5\") " pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:51:35 crc kubenswrapper[4629]: I1211 08:51:35.009522 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/210c22a1-a398-48f1-891f-21fecd9f53d5-apiservice-cert\") pod \"metallb-operator-webhook-server-649886db6b-999qk\" (UID: \"210c22a1-a398-48f1-891f-21fecd9f53d5\") " pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:51:35 crc kubenswrapper[4629]: I1211 08:51:35.009595 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/210c22a1-a398-48f1-891f-21fecd9f53d5-webhook-cert\") pod \"metallb-operator-webhook-server-649886db6b-999qk\" (UID: \"210c22a1-a398-48f1-891f-21fecd9f53d5\") " pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:51:35 crc kubenswrapper[4629]: I1211 08:51:35.027500 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htbtr\" (UniqueName: \"kubernetes.io/projected/210c22a1-a398-48f1-891f-21fecd9f53d5-kube-api-access-htbtr\") pod \"metallb-operator-webhook-server-649886db6b-999qk\" (UID: \"210c22a1-a398-48f1-891f-21fecd9f53d5\") " pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:51:35 crc kubenswrapper[4629]: I1211 08:51:35.043794 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj"] Dec 11 08:51:35 crc kubenswrapper[4629]: W1211 08:51:35.053908 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb51a6d95_cf61_40c5_9d03_92c5c91c16ab.slice/crio-40a9ab89d0c52ded9579ca34eb4bb8da67e3a9883fa218d0e4cf08cb0b2ae20b WatchSource:0}: Error finding container 40a9ab89d0c52ded9579ca34eb4bb8da67e3a9883fa218d0e4cf08cb0b2ae20b: Status 404 returned error can't find the container with id 40a9ab89d0c52ded9579ca34eb4bb8da67e3a9883fa218d0e4cf08cb0b2ae20b Dec 11 08:51:35 crc kubenswrapper[4629]: I1211 08:51:35.112805 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:51:35 crc kubenswrapper[4629]: I1211 08:51:35.821088 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-649886db6b-999qk"] Dec 11 08:51:35 crc kubenswrapper[4629]: I1211 08:51:35.887344 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" event={"ID":"b51a6d95-cf61-40c5-9d03-92c5c91c16ab","Type":"ContainerStarted","Data":"40a9ab89d0c52ded9579ca34eb4bb8da67e3a9883fa218d0e4cf08cb0b2ae20b"} Dec 11 08:51:35 crc kubenswrapper[4629]: I1211 08:51:35.888396 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" event={"ID":"210c22a1-a398-48f1-891f-21fecd9f53d5","Type":"ContainerStarted","Data":"00bc6b11034b6456ac9c944c6d5f0b4765671aa8e1ffce95cfef49a4b02e9306"} Dec 11 08:51:40 crc kubenswrapper[4629]: I1211 08:51:40.930365 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" event={"ID":"b51a6d95-cf61-40c5-9d03-92c5c91c16ab","Type":"ContainerStarted","Data":"1d5701a6507fca455195f8ff6d8cbb41662729945b883c9645fd1e31658c9119"} Dec 11 08:51:40 crc kubenswrapper[4629]: I1211 08:51:40.931633 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:51:40 crc kubenswrapper[4629]: I1211 08:51:40.953140 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" podStartSLOduration=2.035570167 podStartE2EDuration="6.953100295s" podCreationTimestamp="2025-12-11 08:51:34 +0000 UTC" firstStartedPulling="2025-12-11 08:51:35.056671237 +0000 UTC m=+883.122086845" lastFinishedPulling="2025-12-11 08:51:39.974201365 +0000 UTC m=+888.039616973" observedRunningTime="2025-12-11 08:51:40.949685347 +0000 UTC m=+889.015100955" watchObservedRunningTime="2025-12-11 08:51:40.953100295 +0000 UTC m=+889.018515903" Dec 11 08:51:47 crc kubenswrapper[4629]: I1211 08:51:47.063237 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" event={"ID":"210c22a1-a398-48f1-891f-21fecd9f53d5","Type":"ContainerStarted","Data":"e68264605477c266450d678c9afb1ee9cebe238fc0cd3f5138c20dee98acbee4"} Dec 11 08:51:47 crc kubenswrapper[4629]: I1211 08:51:47.063944 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:51:47 crc kubenswrapper[4629]: I1211 08:51:47.124134 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" podStartSLOduration=2.507189765 podStartE2EDuration="13.12411412s" podCreationTimestamp="2025-12-11 08:51:34 +0000 UTC" firstStartedPulling="2025-12-11 08:51:35.83575167 +0000 UTC m=+883.901167278" lastFinishedPulling="2025-12-11 08:51:46.452676025 +0000 UTC m=+894.518091633" observedRunningTime="2025-12-11 08:51:47.119774432 +0000 UTC m=+895.185190040" watchObservedRunningTime="2025-12-11 08:51:47.12411412 +0000 UTC m=+895.189529728" Dec 11 08:52:05 crc kubenswrapper[4629]: I1211 08:52:05.122724 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-649886db6b-999qk" Dec 11 08:52:07 crc kubenswrapper[4629]: I1211 08:52:07.987657 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fgpw6"] Dec 11 08:52:07 crc kubenswrapper[4629]: I1211 08:52:07.992373 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:07 crc kubenswrapper[4629]: I1211 08:52:07.999809 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fgpw6"] Dec 11 08:52:08 crc kubenswrapper[4629]: I1211 08:52:08.136673 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/054f4be9-3fb5-4d4f-baa5-8b743e42255c-utilities\") pod \"community-operators-fgpw6\" (UID: \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\") " pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:08 crc kubenswrapper[4629]: I1211 08:52:08.137104 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvgwv\" (UniqueName: \"kubernetes.io/projected/054f4be9-3fb5-4d4f-baa5-8b743e42255c-kube-api-access-vvgwv\") pod \"community-operators-fgpw6\" (UID: \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\") " pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:08 crc kubenswrapper[4629]: I1211 08:52:08.137126 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/054f4be9-3fb5-4d4f-baa5-8b743e42255c-catalog-content\") pod \"community-operators-fgpw6\" (UID: \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\") " pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:08 crc kubenswrapper[4629]: I1211 08:52:08.238365 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvgwv\" (UniqueName: \"kubernetes.io/projected/054f4be9-3fb5-4d4f-baa5-8b743e42255c-kube-api-access-vvgwv\") pod \"community-operators-fgpw6\" (UID: \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\") " pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:08 crc kubenswrapper[4629]: I1211 08:52:08.238423 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/054f4be9-3fb5-4d4f-baa5-8b743e42255c-catalog-content\") pod \"community-operators-fgpw6\" (UID: \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\") " pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:08 crc kubenswrapper[4629]: I1211 08:52:08.238453 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/054f4be9-3fb5-4d4f-baa5-8b743e42255c-utilities\") pod \"community-operators-fgpw6\" (UID: \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\") " pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:08 crc kubenswrapper[4629]: I1211 08:52:08.239067 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/054f4be9-3fb5-4d4f-baa5-8b743e42255c-catalog-content\") pod \"community-operators-fgpw6\" (UID: \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\") " pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:08 crc kubenswrapper[4629]: I1211 08:52:08.239203 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/054f4be9-3fb5-4d4f-baa5-8b743e42255c-utilities\") pod \"community-operators-fgpw6\" (UID: \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\") " pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:08 crc kubenswrapper[4629]: I1211 08:52:08.277530 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvgwv\" (UniqueName: \"kubernetes.io/projected/054f4be9-3fb5-4d4f-baa5-8b743e42255c-kube-api-access-vvgwv\") pod \"community-operators-fgpw6\" (UID: \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\") " pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:08 crc kubenswrapper[4629]: I1211 08:52:08.321808 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:08 crc kubenswrapper[4629]: I1211 08:52:08.909467 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fgpw6"] Dec 11 08:52:09 crc kubenswrapper[4629]: I1211 08:52:09.182138 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fgpw6" event={"ID":"054f4be9-3fb5-4d4f-baa5-8b743e42255c","Type":"ContainerStarted","Data":"91301b40afe146e91e0b4802951e4c0ee4f0ec44fc195939b51db821640384ed"} Dec 11 08:52:10 crc kubenswrapper[4629]: I1211 08:52:10.188135 4629 generic.go:334] "Generic (PLEG): container finished" podID="054f4be9-3fb5-4d4f-baa5-8b743e42255c" containerID="80844b2ad5680d84d466858faa5c317d79e4abaae4570fca6b76f5b773a1fd71" exitCode=0 Dec 11 08:52:10 crc kubenswrapper[4629]: I1211 08:52:10.188286 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fgpw6" event={"ID":"054f4be9-3fb5-4d4f-baa5-8b743e42255c","Type":"ContainerDied","Data":"80844b2ad5680d84d466858faa5c317d79e4abaae4570fca6b76f5b773a1fd71"} Dec 11 08:52:11 crc kubenswrapper[4629]: I1211 08:52:11.205187 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fgpw6" event={"ID":"054f4be9-3fb5-4d4f-baa5-8b743e42255c","Type":"ContainerStarted","Data":"9bf0ad778f6be4f2b12a8580005fb6967b6cf7498d4b88e86c49f335f5ceb72c"} Dec 11 08:52:12 crc kubenswrapper[4629]: I1211 08:52:12.217164 4629 generic.go:334] "Generic (PLEG): container finished" podID="054f4be9-3fb5-4d4f-baa5-8b743e42255c" containerID="9bf0ad778f6be4f2b12a8580005fb6967b6cf7498d4b88e86c49f335f5ceb72c" exitCode=0 Dec 11 08:52:12 crc kubenswrapper[4629]: I1211 08:52:12.217446 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fgpw6" event={"ID":"054f4be9-3fb5-4d4f-baa5-8b743e42255c","Type":"ContainerDied","Data":"9bf0ad778f6be4f2b12a8580005fb6967b6cf7498d4b88e86c49f335f5ceb72c"} Dec 11 08:52:13 crc kubenswrapper[4629]: I1211 08:52:13.225472 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fgpw6" event={"ID":"054f4be9-3fb5-4d4f-baa5-8b743e42255c","Type":"ContainerStarted","Data":"34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95"} Dec 11 08:52:14 crc kubenswrapper[4629]: I1211 08:52:14.714909 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6f5496ff8f-p5ldj" Dec 11 08:52:14 crc kubenswrapper[4629]: I1211 08:52:14.738162 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fgpw6" podStartSLOduration=4.920266247 podStartE2EDuration="7.738128765s" podCreationTimestamp="2025-12-11 08:52:07 +0000 UTC" firstStartedPulling="2025-12-11 08:52:10.189513816 +0000 UTC m=+918.254929424" lastFinishedPulling="2025-12-11 08:52:13.007376334 +0000 UTC m=+921.072791942" observedRunningTime="2025-12-11 08:52:13.481872055 +0000 UTC m=+921.547287673" watchObservedRunningTime="2025-12-11 08:52:14.738128765 +0000 UTC m=+922.803544383" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.364609 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cfwjh"] Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.365763 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.381650 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cfwjh"] Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.388726 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x66z\" (UniqueName: \"kubernetes.io/projected/77fc0b67-9596-4033-bc05-b0487f495258-kube-api-access-2x66z\") pod \"certified-operators-cfwjh\" (UID: \"77fc0b67-9596-4033-bc05-b0487f495258\") " pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.388778 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77fc0b67-9596-4033-bc05-b0487f495258-utilities\") pod \"certified-operators-cfwjh\" (UID: \"77fc0b67-9596-4033-bc05-b0487f495258\") " pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.388824 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77fc0b67-9596-4033-bc05-b0487f495258-catalog-content\") pod \"certified-operators-cfwjh\" (UID: \"77fc0b67-9596-4033-bc05-b0487f495258\") " pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.490079 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x66z\" (UniqueName: \"kubernetes.io/projected/77fc0b67-9596-4033-bc05-b0487f495258-kube-api-access-2x66z\") pod \"certified-operators-cfwjh\" (UID: \"77fc0b67-9596-4033-bc05-b0487f495258\") " pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.490457 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77fc0b67-9596-4033-bc05-b0487f495258-utilities\") pod \"certified-operators-cfwjh\" (UID: \"77fc0b67-9596-4033-bc05-b0487f495258\") " pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.490511 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77fc0b67-9596-4033-bc05-b0487f495258-catalog-content\") pod \"certified-operators-cfwjh\" (UID: \"77fc0b67-9596-4033-bc05-b0487f495258\") " pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.491092 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77fc0b67-9596-4033-bc05-b0487f495258-catalog-content\") pod \"certified-operators-cfwjh\" (UID: \"77fc0b67-9596-4033-bc05-b0487f495258\") " pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.492093 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77fc0b67-9596-4033-bc05-b0487f495258-utilities\") pod \"certified-operators-cfwjh\" (UID: \"77fc0b67-9596-4033-bc05-b0487f495258\") " pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.531754 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x66z\" (UniqueName: \"kubernetes.io/projected/77fc0b67-9596-4033-bc05-b0487f495258-kube-api-access-2x66z\") pod \"certified-operators-cfwjh\" (UID: \"77fc0b67-9596-4033-bc05-b0487f495258\") " pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.578018 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-t6bzz"] Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.588408 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.598771 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.599342 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.599543 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-b82wb" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.615591 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz"] Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.616494 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.621055 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.639758 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz"] Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.683331 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.698123 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/634a17de-f35e-45c3-ac4e-70b3c1768dac-metrics\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.698198 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c7c31d6-3370-4d14-861b-552a83ce767e-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-tf2sz\" (UID: \"1c7c31d6-3370-4d14-861b-552a83ce767e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.698219 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/634a17de-f35e-45c3-ac4e-70b3c1768dac-frr-conf\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.698252 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/634a17de-f35e-45c3-ac4e-70b3c1768dac-metrics-certs\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.698272 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/634a17de-f35e-45c3-ac4e-70b3c1768dac-frr-startup\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.698292 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5ltp\" (UniqueName: \"kubernetes.io/projected/1c7c31d6-3370-4d14-861b-552a83ce767e-kube-api-access-m5ltp\") pod \"frr-k8s-webhook-server-7784b6fcf-tf2sz\" (UID: \"1c7c31d6-3370-4d14-861b-552a83ce767e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.698324 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/634a17de-f35e-45c3-ac4e-70b3c1768dac-reloader\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.698372 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/634a17de-f35e-45c3-ac4e-70b3c1768dac-frr-sockets\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.698398 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65mcw\" (UniqueName: \"kubernetes.io/projected/634a17de-f35e-45c3-ac4e-70b3c1768dac-kube-api-access-65mcw\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.711011 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-z8mqj"] Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.712172 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-z8mqj" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.726376 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-78sx8" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.726537 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.728151 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.728297 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.733941 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5bddd4b946-ftrtv"] Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.735051 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.737736 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.783954 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-ftrtv"] Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799156 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/634a17de-f35e-45c3-ac4e-70b3c1768dac-reloader\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799224 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdd0751-6dee-4183-b640-85941bc266b4-metrics-certs\") pod \"controller-5bddd4b946-ftrtv\" (UID: \"3fdd0751-6dee-4183-b640-85941bc266b4\") " pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799261 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/634a17de-f35e-45c3-ac4e-70b3c1768dac-frr-sockets\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799300 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65mcw\" (UniqueName: \"kubernetes.io/projected/634a17de-f35e-45c3-ac4e-70b3c1768dac-kube-api-access-65mcw\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799316 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/634a17de-f35e-45c3-ac4e-70b3c1768dac-metrics\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799338 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fdd0751-6dee-4183-b640-85941bc266b4-cert\") pod \"controller-5bddd4b946-ftrtv\" (UID: \"3fdd0751-6dee-4183-b640-85941bc266b4\") " pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799372 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8c79723c-cb8e-4bc9-a088-2881a94dabbc-metallb-excludel2\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799393 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-metrics-certs\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799411 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rfgx\" (UniqueName: \"kubernetes.io/projected/3fdd0751-6dee-4183-b640-85941bc266b4-kube-api-access-2rfgx\") pod \"controller-5bddd4b946-ftrtv\" (UID: \"3fdd0751-6dee-4183-b640-85941bc266b4\") " pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799429 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-memberlist\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799459 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/634a17de-f35e-45c3-ac4e-70b3c1768dac-frr-conf\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799494 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c7c31d6-3370-4d14-861b-552a83ce767e-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-tf2sz\" (UID: \"1c7c31d6-3370-4d14-861b-552a83ce767e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799539 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/634a17de-f35e-45c3-ac4e-70b3c1768dac-metrics-certs\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799556 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5ltp\" (UniqueName: \"kubernetes.io/projected/1c7c31d6-3370-4d14-861b-552a83ce767e-kube-api-access-m5ltp\") pod \"frr-k8s-webhook-server-7784b6fcf-tf2sz\" (UID: \"1c7c31d6-3370-4d14-861b-552a83ce767e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799573 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/634a17de-f35e-45c3-ac4e-70b3c1768dac-frr-startup\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.799606 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b2b6\" (UniqueName: \"kubernetes.io/projected/8c79723c-cb8e-4bc9-a088-2881a94dabbc-kube-api-access-8b2b6\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.800256 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/634a17de-f35e-45c3-ac4e-70b3c1768dac-reloader\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.800334 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/634a17de-f35e-45c3-ac4e-70b3c1768dac-frr-sockets\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: E1211 08:52:15.800442 4629 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 11 08:52:15 crc kubenswrapper[4629]: E1211 08:52:15.800502 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/634a17de-f35e-45c3-ac4e-70b3c1768dac-metrics-certs podName:634a17de-f35e-45c3-ac4e-70b3c1768dac nodeName:}" failed. No retries permitted until 2025-12-11 08:52:16.300479876 +0000 UTC m=+924.365895474 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/634a17de-f35e-45c3-ac4e-70b3c1768dac-metrics-certs") pod "frr-k8s-t6bzz" (UID: "634a17de-f35e-45c3-ac4e-70b3c1768dac") : secret "frr-k8s-certs-secret" not found Dec 11 08:52:15 crc kubenswrapper[4629]: E1211 08:52:15.800535 4629 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 11 08:52:15 crc kubenswrapper[4629]: E1211 08:52:15.800600 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1c7c31d6-3370-4d14-861b-552a83ce767e-cert podName:1c7c31d6-3370-4d14-861b-552a83ce767e nodeName:}" failed. No retries permitted until 2025-12-11 08:52:16.300568029 +0000 UTC m=+924.365983717 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1c7c31d6-3370-4d14-861b-552a83ce767e-cert") pod "frr-k8s-webhook-server-7784b6fcf-tf2sz" (UID: "1c7c31d6-3370-4d14-861b-552a83ce767e") : secret "frr-k8s-webhook-server-cert" not found Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.802037 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/634a17de-f35e-45c3-ac4e-70b3c1768dac-frr-conf\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.802286 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/634a17de-f35e-45c3-ac4e-70b3c1768dac-metrics\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.803318 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/634a17de-f35e-45c3-ac4e-70b3c1768dac-frr-startup\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.833751 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65mcw\" (UniqueName: \"kubernetes.io/projected/634a17de-f35e-45c3-ac4e-70b3c1768dac-kube-api-access-65mcw\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.914217 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rfgx\" (UniqueName: \"kubernetes.io/projected/3fdd0751-6dee-4183-b640-85941bc266b4-kube-api-access-2rfgx\") pod \"controller-5bddd4b946-ftrtv\" (UID: \"3fdd0751-6dee-4183-b640-85941bc266b4\") " pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.914278 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-memberlist\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.914399 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b2b6\" (UniqueName: \"kubernetes.io/projected/8c79723c-cb8e-4bc9-a088-2881a94dabbc-kube-api-access-8b2b6\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.914474 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdd0751-6dee-4183-b640-85941bc266b4-metrics-certs\") pod \"controller-5bddd4b946-ftrtv\" (UID: \"3fdd0751-6dee-4183-b640-85941bc266b4\") " pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.914725 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fdd0751-6dee-4183-b640-85941bc266b4-cert\") pod \"controller-5bddd4b946-ftrtv\" (UID: \"3fdd0751-6dee-4183-b640-85941bc266b4\") " pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.914754 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8c79723c-cb8e-4bc9-a088-2881a94dabbc-metallb-excludel2\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.914793 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-metrics-certs\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:15 crc kubenswrapper[4629]: E1211 08:52:15.914923 4629 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 11 08:52:15 crc kubenswrapper[4629]: E1211 08:52:15.914988 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-metrics-certs podName:8c79723c-cb8e-4bc9-a088-2881a94dabbc nodeName:}" failed. No retries permitted until 2025-12-11 08:52:16.414972832 +0000 UTC m=+924.480388440 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-metrics-certs") pod "speaker-z8mqj" (UID: "8c79723c-cb8e-4bc9-a088-2881a94dabbc") : secret "speaker-certs-secret" not found Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.914468 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5ltp\" (UniqueName: \"kubernetes.io/projected/1c7c31d6-3370-4d14-861b-552a83ce767e-kube-api-access-m5ltp\") pod \"frr-k8s-webhook-server-7784b6fcf-tf2sz\" (UID: \"1c7c31d6-3370-4d14-861b-552a83ce767e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" Dec 11 08:52:15 crc kubenswrapper[4629]: E1211 08:52:15.915201 4629 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 11 08:52:15 crc kubenswrapper[4629]: E1211 08:52:15.915304 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fdd0751-6dee-4183-b640-85941bc266b4-metrics-certs podName:3fdd0751-6dee-4183-b640-85941bc266b4 nodeName:}" failed. No retries permitted until 2025-12-11 08:52:16.415285401 +0000 UTC m=+924.480701009 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fdd0751-6dee-4183-b640-85941bc266b4-metrics-certs") pod "controller-5bddd4b946-ftrtv" (UID: "3fdd0751-6dee-4183-b640-85941bc266b4") : secret "controller-certs-secret" not found Dec 11 08:52:15 crc kubenswrapper[4629]: E1211 08:52:15.915478 4629 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 11 08:52:15 crc kubenswrapper[4629]: E1211 08:52:15.915601 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-memberlist podName:8c79723c-cb8e-4bc9-a088-2881a94dabbc nodeName:}" failed. No retries permitted until 2025-12-11 08:52:16.415579381 +0000 UTC m=+924.480994989 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-memberlist") pod "speaker-z8mqj" (UID: "8c79723c-cb8e-4bc9-a088-2881a94dabbc") : secret "metallb-memberlist" not found Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.916227 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8c79723c-cb8e-4bc9-a088-2881a94dabbc-metallb-excludel2\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.927137 4629 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.939814 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fdd0751-6dee-4183-b640-85941bc266b4-cert\") pod \"controller-5bddd4b946-ftrtv\" (UID: \"3fdd0751-6dee-4183-b640-85941bc266b4\") " pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.963657 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b2b6\" (UniqueName: \"kubernetes.io/projected/8c79723c-cb8e-4bc9-a088-2881a94dabbc-kube-api-access-8b2b6\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:15 crc kubenswrapper[4629]: I1211 08:52:15.965763 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rfgx\" (UniqueName: \"kubernetes.io/projected/3fdd0751-6dee-4183-b640-85941bc266b4-kube-api-access-2rfgx\") pod \"controller-5bddd4b946-ftrtv\" (UID: \"3fdd0751-6dee-4183-b640-85941bc266b4\") " pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.331782 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c7c31d6-3370-4d14-861b-552a83ce767e-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-tf2sz\" (UID: \"1c7c31d6-3370-4d14-861b-552a83ce767e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.332188 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/634a17de-f35e-45c3-ac4e-70b3c1768dac-metrics-certs\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.336096 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c7c31d6-3370-4d14-861b-552a83ce767e-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-tf2sz\" (UID: \"1c7c31d6-3370-4d14-861b-552a83ce767e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.337309 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/634a17de-f35e-45c3-ac4e-70b3c1768dac-metrics-certs\") pod \"frr-k8s-t6bzz\" (UID: \"634a17de-f35e-45c3-ac4e-70b3c1768dac\") " pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.385422 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cfwjh"] Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.434574 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdd0751-6dee-4183-b640-85941bc266b4-metrics-certs\") pod \"controller-5bddd4b946-ftrtv\" (UID: \"3fdd0751-6dee-4183-b640-85941bc266b4\") " pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.434673 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-metrics-certs\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.434705 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-memberlist\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:16 crc kubenswrapper[4629]: E1211 08:52:16.434830 4629 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 11 08:52:16 crc kubenswrapper[4629]: E1211 08:52:16.434895 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-memberlist podName:8c79723c-cb8e-4bc9-a088-2881a94dabbc nodeName:}" failed. No retries permitted until 2025-12-11 08:52:17.434880808 +0000 UTC m=+925.500296416 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-memberlist") pod "speaker-z8mqj" (UID: "8c79723c-cb8e-4bc9-a088-2881a94dabbc") : secret "metallb-memberlist" not found Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.444311 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-metrics-certs\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.446390 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdd0751-6dee-4183-b640-85941bc266b4-metrics-certs\") pod \"controller-5bddd4b946-ftrtv\" (UID: \"3fdd0751-6dee-4183-b640-85941bc266b4\") " pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.531221 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.549747 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.720111 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:16 crc kubenswrapper[4629]: I1211 08:52:16.990795 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz"] Dec 11 08:52:17 crc kubenswrapper[4629]: W1211 08:52:17.012913 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c7c31d6_3370_4d14_861b_552a83ce767e.slice/crio-aa0efd0d0b9310dac1f2d63f3213d0d290232b91053fcc78a8eb09d0b496de91 WatchSource:0}: Error finding container aa0efd0d0b9310dac1f2d63f3213d0d290232b91053fcc78a8eb09d0b496de91: Status 404 returned error can't find the container with id aa0efd0d0b9310dac1f2d63f3213d0d290232b91053fcc78a8eb09d0b496de91 Dec 11 08:52:17 crc kubenswrapper[4629]: I1211 08:52:17.225514 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-ftrtv"] Dec 11 08:52:17 crc kubenswrapper[4629]: W1211 08:52:17.228460 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fdd0751_6dee_4183_b640_85941bc266b4.slice/crio-dbad01c671df52663fb5ccd4417bae78eb16908c63da2c71ce00c23b3811f09f WatchSource:0}: Error finding container dbad01c671df52663fb5ccd4417bae78eb16908c63da2c71ce00c23b3811f09f: Status 404 returned error can't find the container with id dbad01c671df52663fb5ccd4417bae78eb16908c63da2c71ce00c23b3811f09f Dec 11 08:52:17 crc kubenswrapper[4629]: I1211 08:52:17.256601 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" event={"ID":"1c7c31d6-3370-4d14-861b-552a83ce767e","Type":"ContainerStarted","Data":"aa0efd0d0b9310dac1f2d63f3213d0d290232b91053fcc78a8eb09d0b496de91"} Dec 11 08:52:17 crc kubenswrapper[4629]: I1211 08:52:17.258136 4629 generic.go:334] "Generic (PLEG): container finished" podID="77fc0b67-9596-4033-bc05-b0487f495258" containerID="a02eaf577326d76bb0f84e36e0cf7283d4b062ec9ca5509c2475abb4c6f400cf" exitCode=0 Dec 11 08:52:17 crc kubenswrapper[4629]: I1211 08:52:17.258384 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfwjh" event={"ID":"77fc0b67-9596-4033-bc05-b0487f495258","Type":"ContainerDied","Data":"a02eaf577326d76bb0f84e36e0cf7283d4b062ec9ca5509c2475abb4c6f400cf"} Dec 11 08:52:17 crc kubenswrapper[4629]: I1211 08:52:17.258457 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfwjh" event={"ID":"77fc0b67-9596-4033-bc05-b0487f495258","Type":"ContainerStarted","Data":"106a2dc076cc94b77a6584bd126a054de586c608e8c433a5a5f94e07510d56a3"} Dec 11 08:52:17 crc kubenswrapper[4629]: I1211 08:52:17.260653 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-ftrtv" event={"ID":"3fdd0751-6dee-4183-b640-85941bc266b4","Type":"ContainerStarted","Data":"dbad01c671df52663fb5ccd4417bae78eb16908c63da2c71ce00c23b3811f09f"} Dec 11 08:52:17 crc kubenswrapper[4629]: I1211 08:52:17.267367 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t6bzz" event={"ID":"634a17de-f35e-45c3-ac4e-70b3c1768dac","Type":"ContainerStarted","Data":"81a9a515dcc6c6437b9ea8eaf4a1c0ef09a4d003e94031982a3f58f8a95bae22"} Dec 11 08:52:17 crc kubenswrapper[4629]: I1211 08:52:17.494537 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-memberlist\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:17 crc kubenswrapper[4629]: E1211 08:52:17.494677 4629 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 11 08:52:17 crc kubenswrapper[4629]: E1211 08:52:17.494762 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-memberlist podName:8c79723c-cb8e-4bc9-a088-2881a94dabbc nodeName:}" failed. No retries permitted until 2025-12-11 08:52:19.494745061 +0000 UTC m=+927.560160669 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-memberlist") pod "speaker-z8mqj" (UID: "8c79723c-cb8e-4bc9-a088-2881a94dabbc") : secret "metallb-memberlist" not found Dec 11 08:52:18 crc kubenswrapper[4629]: I1211 08:52:18.322613 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:18 crc kubenswrapper[4629]: I1211 08:52:18.322941 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:18 crc kubenswrapper[4629]: I1211 08:52:18.334720 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-ftrtv" event={"ID":"3fdd0751-6dee-4183-b640-85941bc266b4","Type":"ContainerStarted","Data":"86ee793e9bbfdd756623ca9d0a3781a3900ca76cbf1df24e493b3325209eff7f"} Dec 11 08:52:18 crc kubenswrapper[4629]: I1211 08:52:18.334808 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-ftrtv" event={"ID":"3fdd0751-6dee-4183-b640-85941bc266b4","Type":"ContainerStarted","Data":"a88cc049336b74e7b85f9a37c53f1446ba76bcecade74cb3a927c1b0727088a5"} Dec 11 08:52:18 crc kubenswrapper[4629]: I1211 08:52:18.335921 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:18 crc kubenswrapper[4629]: I1211 08:52:18.384373 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5bddd4b946-ftrtv" podStartSLOduration=3.384348427 podStartE2EDuration="3.384348427s" podCreationTimestamp="2025-12-11 08:52:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:52:18.383300823 +0000 UTC m=+926.448716421" watchObservedRunningTime="2025-12-11 08:52:18.384348427 +0000 UTC m=+926.449764035" Dec 11 08:52:18 crc kubenswrapper[4629]: I1211 08:52:18.454917 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:19 crc kubenswrapper[4629]: I1211 08:52:19.413633 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:19 crc kubenswrapper[4629]: I1211 08:52:19.592509 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-memberlist\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:19 crc kubenswrapper[4629]: I1211 08:52:19.613136 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8c79723c-cb8e-4bc9-a088-2881a94dabbc-memberlist\") pod \"speaker-z8mqj\" (UID: \"8c79723c-cb8e-4bc9-a088-2881a94dabbc\") " pod="metallb-system/speaker-z8mqj" Dec 11 08:52:19 crc kubenswrapper[4629]: I1211 08:52:19.657464 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-z8mqj" Dec 11 08:52:20 crc kubenswrapper[4629]: I1211 08:52:20.470917 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-z8mqj" event={"ID":"8c79723c-cb8e-4bc9-a088-2881a94dabbc","Type":"ContainerStarted","Data":"e33b87e0402c35c4d350d5e9c951d846599de4b85df69c213b6d4a5abede1ab4"} Dec 11 08:52:20 crc kubenswrapper[4629]: I1211 08:52:20.471294 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-z8mqj" event={"ID":"8c79723c-cb8e-4bc9-a088-2881a94dabbc","Type":"ContainerStarted","Data":"a07ba1fbcdd28e66e00226f66568bede942ae62f43732dfe0f3e0e57b7cc69c5"} Dec 11 08:52:20 crc kubenswrapper[4629]: I1211 08:52:20.486665 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:52:20 crc kubenswrapper[4629]: I1211 08:52:20.486747 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:52:20 crc kubenswrapper[4629]: I1211 08:52:20.564096 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fgpw6"] Dec 11 08:52:21 crc kubenswrapper[4629]: I1211 08:52:21.476477 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fgpw6" podUID="054f4be9-3fb5-4d4f-baa5-8b743e42255c" containerName="registry-server" containerID="cri-o://34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95" gracePeriod=2 Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.213210 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.518700 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/054f4be9-3fb5-4d4f-baa5-8b743e42255c-utilities\") pod \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\" (UID: \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\") " Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.519084 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/054f4be9-3fb5-4d4f-baa5-8b743e42255c-catalog-content\") pod \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\" (UID: \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\") " Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.519255 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvgwv\" (UniqueName: \"kubernetes.io/projected/054f4be9-3fb5-4d4f-baa5-8b743e42255c-kube-api-access-vvgwv\") pod \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\" (UID: \"054f4be9-3fb5-4d4f-baa5-8b743e42255c\") " Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.520315 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/054f4be9-3fb5-4d4f-baa5-8b743e42255c-utilities" (OuterVolumeSpecName: "utilities") pod "054f4be9-3fb5-4d4f-baa5-8b743e42255c" (UID: "054f4be9-3fb5-4d4f-baa5-8b743e42255c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.541156 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/054f4be9-3fb5-4d4f-baa5-8b743e42255c-kube-api-access-vvgwv" (OuterVolumeSpecName: "kube-api-access-vvgwv") pod "054f4be9-3fb5-4d4f-baa5-8b743e42255c" (UID: "054f4be9-3fb5-4d4f-baa5-8b743e42255c"). InnerVolumeSpecName "kube-api-access-vvgwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.549326 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-z8mqj" event={"ID":"8c79723c-cb8e-4bc9-a088-2881a94dabbc","Type":"ContainerStarted","Data":"05eeb3fcf600514ffb3d4111e5a6781ed1702bbd85c3ad57bb56b7abf7932bca"} Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.552688 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-z8mqj" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.563293 4629 generic.go:334] "Generic (PLEG): container finished" podID="054f4be9-3fb5-4d4f-baa5-8b743e42255c" containerID="34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95" exitCode=0 Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.563374 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fgpw6" event={"ID":"054f4be9-3fb5-4d4f-baa5-8b743e42255c","Type":"ContainerDied","Data":"34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95"} Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.563422 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fgpw6" event={"ID":"054f4be9-3fb5-4d4f-baa5-8b743e42255c","Type":"ContainerDied","Data":"91301b40afe146e91e0b4802951e4c0ee4f0ec44fc195939b51db821640384ed"} Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.563442 4629 scope.go:117] "RemoveContainer" containerID="34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.563624 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fgpw6" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.574956 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfwjh" event={"ID":"77fc0b67-9596-4033-bc05-b0487f495258","Type":"ContainerStarted","Data":"4bf463221610df4c3f173c3e9d1888c75e5fbce5692989adca75dbd3d0e529a7"} Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.578744 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-z8mqj" podStartSLOduration=7.578716852 podStartE2EDuration="7.578716852s" podCreationTimestamp="2025-12-11 08:52:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:52:22.572016931 +0000 UTC m=+930.637432529" watchObservedRunningTime="2025-12-11 08:52:22.578716852 +0000 UTC m=+930.644132460" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.618542 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/054f4be9-3fb5-4d4f-baa5-8b743e42255c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "054f4be9-3fb5-4d4f-baa5-8b743e42255c" (UID: "054f4be9-3fb5-4d4f-baa5-8b743e42255c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.620732 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvgwv\" (UniqueName: \"kubernetes.io/projected/054f4be9-3fb5-4d4f-baa5-8b743e42255c-kube-api-access-vvgwv\") on node \"crc\" DevicePath \"\"" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.620777 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/054f4be9-3fb5-4d4f-baa5-8b743e42255c-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.620791 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/054f4be9-3fb5-4d4f-baa5-8b743e42255c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.634711 4629 scope.go:117] "RemoveContainer" containerID="9bf0ad778f6be4f2b12a8580005fb6967b6cf7498d4b88e86c49f335f5ceb72c" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.669131 4629 scope.go:117] "RemoveContainer" containerID="80844b2ad5680d84d466858faa5c317d79e4abaae4570fca6b76f5b773a1fd71" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.713696 4629 scope.go:117] "RemoveContainer" containerID="34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95" Dec 11 08:52:22 crc kubenswrapper[4629]: E1211 08:52:22.718936 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95\": container with ID starting with 34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95 not found: ID does not exist" containerID="34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.718978 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95"} err="failed to get container status \"34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95\": rpc error: code = NotFound desc = could not find container \"34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95\": container with ID starting with 34901abf16f6f815d8f78dc38fa463315e0f98db5d0614ec059ec6b1c6bb4b95 not found: ID does not exist" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.719008 4629 scope.go:117] "RemoveContainer" containerID="9bf0ad778f6be4f2b12a8580005fb6967b6cf7498d4b88e86c49f335f5ceb72c" Dec 11 08:52:22 crc kubenswrapper[4629]: E1211 08:52:22.723012 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bf0ad778f6be4f2b12a8580005fb6967b6cf7498d4b88e86c49f335f5ceb72c\": container with ID starting with 9bf0ad778f6be4f2b12a8580005fb6967b6cf7498d4b88e86c49f335f5ceb72c not found: ID does not exist" containerID="9bf0ad778f6be4f2b12a8580005fb6967b6cf7498d4b88e86c49f335f5ceb72c" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.723057 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bf0ad778f6be4f2b12a8580005fb6967b6cf7498d4b88e86c49f335f5ceb72c"} err="failed to get container status \"9bf0ad778f6be4f2b12a8580005fb6967b6cf7498d4b88e86c49f335f5ceb72c\": rpc error: code = NotFound desc = could not find container \"9bf0ad778f6be4f2b12a8580005fb6967b6cf7498d4b88e86c49f335f5ceb72c\": container with ID starting with 9bf0ad778f6be4f2b12a8580005fb6967b6cf7498d4b88e86c49f335f5ceb72c not found: ID does not exist" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.723083 4629 scope.go:117] "RemoveContainer" containerID="80844b2ad5680d84d466858faa5c317d79e4abaae4570fca6b76f5b773a1fd71" Dec 11 08:52:22 crc kubenswrapper[4629]: E1211 08:52:22.726978 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80844b2ad5680d84d466858faa5c317d79e4abaae4570fca6b76f5b773a1fd71\": container with ID starting with 80844b2ad5680d84d466858faa5c317d79e4abaae4570fca6b76f5b773a1fd71 not found: ID does not exist" containerID="80844b2ad5680d84d466858faa5c317d79e4abaae4570fca6b76f5b773a1fd71" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.727015 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80844b2ad5680d84d466858faa5c317d79e4abaae4570fca6b76f5b773a1fd71"} err="failed to get container status \"80844b2ad5680d84d466858faa5c317d79e4abaae4570fca6b76f5b773a1fd71\": rpc error: code = NotFound desc = could not find container \"80844b2ad5680d84d466858faa5c317d79e4abaae4570fca6b76f5b773a1fd71\": container with ID starting with 80844b2ad5680d84d466858faa5c317d79e4abaae4570fca6b76f5b773a1fd71 not found: ID does not exist" Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.903414 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fgpw6"] Dec 11 08:52:22 crc kubenswrapper[4629]: I1211 08:52:22.909156 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fgpw6"] Dec 11 08:52:24 crc kubenswrapper[4629]: I1211 08:52:24.209350 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="054f4be9-3fb5-4d4f-baa5-8b743e42255c" path="/var/lib/kubelet/pods/054f4be9-3fb5-4d4f-baa5-8b743e42255c/volumes" Dec 11 08:52:24 crc kubenswrapper[4629]: I1211 08:52:24.588418 4629 generic.go:334] "Generic (PLEG): container finished" podID="77fc0b67-9596-4033-bc05-b0487f495258" containerID="4bf463221610df4c3f173c3e9d1888c75e5fbce5692989adca75dbd3d0e529a7" exitCode=0 Dec 11 08:52:24 crc kubenswrapper[4629]: I1211 08:52:24.588457 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfwjh" event={"ID":"77fc0b67-9596-4033-bc05-b0487f495258","Type":"ContainerDied","Data":"4bf463221610df4c3f173c3e9d1888c75e5fbce5692989adca75dbd3d0e529a7"} Dec 11 08:52:25 crc kubenswrapper[4629]: I1211 08:52:25.608056 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfwjh" event={"ID":"77fc0b67-9596-4033-bc05-b0487f495258","Type":"ContainerStarted","Data":"ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f"} Dec 11 08:52:25 crc kubenswrapper[4629]: I1211 08:52:25.637506 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cfwjh" podStartSLOduration=2.739487326 podStartE2EDuration="10.637483466s" podCreationTimestamp="2025-12-11 08:52:15 +0000 UTC" firstStartedPulling="2025-12-11 08:52:17.259278977 +0000 UTC m=+925.324694585" lastFinishedPulling="2025-12-11 08:52:25.157275117 +0000 UTC m=+933.222690725" observedRunningTime="2025-12-11 08:52:25.63442406 +0000 UTC m=+933.699839688" watchObservedRunningTime="2025-12-11 08:52:25.637483466 +0000 UTC m=+933.702899074" Dec 11 08:52:25 crc kubenswrapper[4629]: I1211 08:52:25.685288 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:25 crc kubenswrapper[4629]: I1211 08:52:25.685334 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:26 crc kubenswrapper[4629]: I1211 08:52:26.808507 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-cfwjh" podUID="77fc0b67-9596-4033-bc05-b0487f495258" containerName="registry-server" probeResult="failure" output=< Dec 11 08:52:26 crc kubenswrapper[4629]: timeout: failed to connect service ":50051" within 1s Dec 11 08:52:26 crc kubenswrapper[4629]: > Dec 11 08:52:34 crc kubenswrapper[4629]: I1211 08:52:34.786968 4629 generic.go:334] "Generic (PLEG): container finished" podID="634a17de-f35e-45c3-ac4e-70b3c1768dac" containerID="fb5bb40d9d11e0f8d6829de218da12db2e541bd41e93b1afdb510324f21194fb" exitCode=0 Dec 11 08:52:34 crc kubenswrapper[4629]: I1211 08:52:34.787108 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t6bzz" event={"ID":"634a17de-f35e-45c3-ac4e-70b3c1768dac","Type":"ContainerDied","Data":"fb5bb40d9d11e0f8d6829de218da12db2e541bd41e93b1afdb510324f21194fb"} Dec 11 08:52:34 crc kubenswrapper[4629]: I1211 08:52:34.789155 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" event={"ID":"1c7c31d6-3370-4d14-861b-552a83ce767e","Type":"ContainerStarted","Data":"f6e0fe82d1301371c31901febcb833e93ce3738b05c94a92cf9addcb7efb2591"} Dec 11 08:52:34 crc kubenswrapper[4629]: I1211 08:52:34.789323 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" Dec 11 08:52:34 crc kubenswrapper[4629]: I1211 08:52:34.872889 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" podStartSLOduration=3.230926479 podStartE2EDuration="19.872832492s" podCreationTimestamp="2025-12-11 08:52:15 +0000 UTC" firstStartedPulling="2025-12-11 08:52:17.018217186 +0000 UTC m=+925.083632794" lastFinishedPulling="2025-12-11 08:52:33.660123189 +0000 UTC m=+941.725538807" observedRunningTime="2025-12-11 08:52:34.866938227 +0000 UTC m=+942.932353845" watchObservedRunningTime="2025-12-11 08:52:34.872832492 +0000 UTC m=+942.938248100" Dec 11 08:52:35 crc kubenswrapper[4629]: I1211 08:52:35.778673 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:35 crc kubenswrapper[4629]: I1211 08:52:35.808702 4629 generic.go:334] "Generic (PLEG): container finished" podID="634a17de-f35e-45c3-ac4e-70b3c1768dac" containerID="b1d5119a71c3f5ba3460ebe2451e5f0ce28484abe3a9554d4a0d97cf9cf928c6" exitCode=0 Dec 11 08:52:35 crc kubenswrapper[4629]: I1211 08:52:35.808805 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t6bzz" event={"ID":"634a17de-f35e-45c3-ac4e-70b3c1768dac","Type":"ContainerDied","Data":"b1d5119a71c3f5ba3460ebe2451e5f0ce28484abe3a9554d4a0d97cf9cf928c6"} Dec 11 08:52:35 crc kubenswrapper[4629]: I1211 08:52:35.829623 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:36 crc kubenswrapper[4629]: I1211 08:52:36.014881 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cfwjh"] Dec 11 08:52:36 crc kubenswrapper[4629]: I1211 08:52:36.736047 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5bddd4b946-ftrtv" Dec 11 08:52:36 crc kubenswrapper[4629]: I1211 08:52:36.816475 4629 generic.go:334] "Generic (PLEG): container finished" podID="634a17de-f35e-45c3-ac4e-70b3c1768dac" containerID="0fd4cd74a91079f23c99851d0281eb74c8a0638a115408a70341bba003e960bb" exitCode=0 Dec 11 08:52:36 crc kubenswrapper[4629]: I1211 08:52:36.816544 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t6bzz" event={"ID":"634a17de-f35e-45c3-ac4e-70b3c1768dac","Type":"ContainerDied","Data":"0fd4cd74a91079f23c99851d0281eb74c8a0638a115408a70341bba003e960bb"} Dec 11 08:52:36 crc kubenswrapper[4629]: I1211 08:52:36.816704 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cfwjh" podUID="77fc0b67-9596-4033-bc05-b0487f495258" containerName="registry-server" containerID="cri-o://ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f" gracePeriod=2 Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.290774 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.413721 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77fc0b67-9596-4033-bc05-b0487f495258-catalog-content\") pod \"77fc0b67-9596-4033-bc05-b0487f495258\" (UID: \"77fc0b67-9596-4033-bc05-b0487f495258\") " Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.413858 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x66z\" (UniqueName: \"kubernetes.io/projected/77fc0b67-9596-4033-bc05-b0487f495258-kube-api-access-2x66z\") pod \"77fc0b67-9596-4033-bc05-b0487f495258\" (UID: \"77fc0b67-9596-4033-bc05-b0487f495258\") " Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.413881 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77fc0b67-9596-4033-bc05-b0487f495258-utilities\") pod \"77fc0b67-9596-4033-bc05-b0487f495258\" (UID: \"77fc0b67-9596-4033-bc05-b0487f495258\") " Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.414997 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77fc0b67-9596-4033-bc05-b0487f495258-utilities" (OuterVolumeSpecName: "utilities") pod "77fc0b67-9596-4033-bc05-b0487f495258" (UID: "77fc0b67-9596-4033-bc05-b0487f495258"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.419973 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77fc0b67-9596-4033-bc05-b0487f495258-kube-api-access-2x66z" (OuterVolumeSpecName: "kube-api-access-2x66z") pod "77fc0b67-9596-4033-bc05-b0487f495258" (UID: "77fc0b67-9596-4033-bc05-b0487f495258"). InnerVolumeSpecName "kube-api-access-2x66z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.472614 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77fc0b67-9596-4033-bc05-b0487f495258-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77fc0b67-9596-4033-bc05-b0487f495258" (UID: "77fc0b67-9596-4033-bc05-b0487f495258"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.522372 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x66z\" (UniqueName: \"kubernetes.io/projected/77fc0b67-9596-4033-bc05-b0487f495258-kube-api-access-2x66z\") on node \"crc\" DevicePath \"\"" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.522403 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77fc0b67-9596-4033-bc05-b0487f495258-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.522415 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77fc0b67-9596-4033-bc05-b0487f495258-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.827738 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t6bzz" event={"ID":"634a17de-f35e-45c3-ac4e-70b3c1768dac","Type":"ContainerStarted","Data":"20cdc8d43f2a0eb9194ea76f0b2568ae6cfc745238d4d3a30a7439ccb0d40984"} Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.828130 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t6bzz" event={"ID":"634a17de-f35e-45c3-ac4e-70b3c1768dac","Type":"ContainerStarted","Data":"e7e9ad6fbd103f73eaab00467464517a521c9a583cc5e84152d63faafafad81e"} Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.828141 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t6bzz" event={"ID":"634a17de-f35e-45c3-ac4e-70b3c1768dac","Type":"ContainerStarted","Data":"b1928bfc2847f66715e4d5f27e075382c3d15b369a9d52716e624a575360d403"} Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.828149 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t6bzz" event={"ID":"634a17de-f35e-45c3-ac4e-70b3c1768dac","Type":"ContainerStarted","Data":"b7cffa130e12b8916136ec1f54c5270820606b11a298772e4097eec87cd93cc0"} Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.828158 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t6bzz" event={"ID":"634a17de-f35e-45c3-ac4e-70b3c1768dac","Type":"ContainerStarted","Data":"55f15ddb3cf8ceb381453b8a9a5cbd35fe101b7aced0a773fd99b29a491210fe"} Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.832452 4629 generic.go:334] "Generic (PLEG): container finished" podID="77fc0b67-9596-4033-bc05-b0487f495258" containerID="ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f" exitCode=0 Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.832503 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfwjh" event={"ID":"77fc0b67-9596-4033-bc05-b0487f495258","Type":"ContainerDied","Data":"ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f"} Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.832536 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfwjh" event={"ID":"77fc0b67-9596-4033-bc05-b0487f495258","Type":"ContainerDied","Data":"106a2dc076cc94b77a6584bd126a054de586c608e8c433a5a5f94e07510d56a3"} Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.832558 4629 scope.go:117] "RemoveContainer" containerID="ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.832701 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cfwjh" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.857177 4629 scope.go:117] "RemoveContainer" containerID="4bf463221610df4c3f173c3e9d1888c75e5fbce5692989adca75dbd3d0e529a7" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.873467 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cfwjh"] Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.881079 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cfwjh"] Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.894021 4629 scope.go:117] "RemoveContainer" containerID="a02eaf577326d76bb0f84e36e0cf7283d4b062ec9ca5509c2475abb4c6f400cf" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.930591 4629 scope.go:117] "RemoveContainer" containerID="ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f" Dec 11 08:52:37 crc kubenswrapper[4629]: E1211 08:52:37.933369 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f\": container with ID starting with ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f not found: ID does not exist" containerID="ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.933416 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f"} err="failed to get container status \"ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f\": rpc error: code = NotFound desc = could not find container \"ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f\": container with ID starting with ae597419020a3bacd288fc520b244ab71814495f85daeabc6220fbdade3f225f not found: ID does not exist" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.933446 4629 scope.go:117] "RemoveContainer" containerID="4bf463221610df4c3f173c3e9d1888c75e5fbce5692989adca75dbd3d0e529a7" Dec 11 08:52:37 crc kubenswrapper[4629]: E1211 08:52:37.934080 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bf463221610df4c3f173c3e9d1888c75e5fbce5692989adca75dbd3d0e529a7\": container with ID starting with 4bf463221610df4c3f173c3e9d1888c75e5fbce5692989adca75dbd3d0e529a7 not found: ID does not exist" containerID="4bf463221610df4c3f173c3e9d1888c75e5fbce5692989adca75dbd3d0e529a7" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.934147 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bf463221610df4c3f173c3e9d1888c75e5fbce5692989adca75dbd3d0e529a7"} err="failed to get container status \"4bf463221610df4c3f173c3e9d1888c75e5fbce5692989adca75dbd3d0e529a7\": rpc error: code = NotFound desc = could not find container \"4bf463221610df4c3f173c3e9d1888c75e5fbce5692989adca75dbd3d0e529a7\": container with ID starting with 4bf463221610df4c3f173c3e9d1888c75e5fbce5692989adca75dbd3d0e529a7 not found: ID does not exist" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.934183 4629 scope.go:117] "RemoveContainer" containerID="a02eaf577326d76bb0f84e36e0cf7283d4b062ec9ca5509c2475abb4c6f400cf" Dec 11 08:52:37 crc kubenswrapper[4629]: E1211 08:52:37.934610 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a02eaf577326d76bb0f84e36e0cf7283d4b062ec9ca5509c2475abb4c6f400cf\": container with ID starting with a02eaf577326d76bb0f84e36e0cf7283d4b062ec9ca5509c2475abb4c6f400cf not found: ID does not exist" containerID="a02eaf577326d76bb0f84e36e0cf7283d4b062ec9ca5509c2475abb4c6f400cf" Dec 11 08:52:37 crc kubenswrapper[4629]: I1211 08:52:37.934657 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a02eaf577326d76bb0f84e36e0cf7283d4b062ec9ca5509c2475abb4c6f400cf"} err="failed to get container status \"a02eaf577326d76bb0f84e36e0cf7283d4b062ec9ca5509c2475abb4c6f400cf\": rpc error: code = NotFound desc = could not find container \"a02eaf577326d76bb0f84e36e0cf7283d4b062ec9ca5509c2475abb4c6f400cf\": container with ID starting with a02eaf577326d76bb0f84e36e0cf7283d4b062ec9ca5509c2475abb4c6f400cf not found: ID does not exist" Dec 11 08:52:38 crc kubenswrapper[4629]: I1211 08:52:38.217012 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77fc0b67-9596-4033-bc05-b0487f495258" path="/var/lib/kubelet/pods/77fc0b67-9596-4033-bc05-b0487f495258/volumes" Dec 11 08:52:38 crc kubenswrapper[4629]: I1211 08:52:38.841836 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t6bzz" event={"ID":"634a17de-f35e-45c3-ac4e-70b3c1768dac","Type":"ContainerStarted","Data":"a75819dc85595b615b2b975c53b19e696bce8ff9bffb057bf912f275a72819d3"} Dec 11 08:52:38 crc kubenswrapper[4629]: I1211 08:52:38.842057 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:38 crc kubenswrapper[4629]: I1211 08:52:38.864044 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-t6bzz" podStartSLOduration=7.415870839 podStartE2EDuration="23.864030908s" podCreationTimestamp="2025-12-11 08:52:15 +0000 UTC" firstStartedPulling="2025-12-11 08:52:17.148363773 +0000 UTC m=+925.213779381" lastFinishedPulling="2025-12-11 08:52:33.596523842 +0000 UTC m=+941.661939450" observedRunningTime="2025-12-11 08:52:38.862445007 +0000 UTC m=+946.927860615" watchObservedRunningTime="2025-12-11 08:52:38.864030908 +0000 UTC m=+946.929446516" Dec 11 08:52:39 crc kubenswrapper[4629]: I1211 08:52:39.668136 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-z8mqj" Dec 11 08:52:41 crc kubenswrapper[4629]: I1211 08:52:41.531716 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:41 crc kubenswrapper[4629]: I1211 08:52:41.610168 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.477059 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2nxg4"] Dec 11 08:52:42 crc kubenswrapper[4629]: E1211 08:52:42.477616 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77fc0b67-9596-4033-bc05-b0487f495258" containerName="registry-server" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.477633 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="77fc0b67-9596-4033-bc05-b0487f495258" containerName="registry-server" Dec 11 08:52:42 crc kubenswrapper[4629]: E1211 08:52:42.477647 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="054f4be9-3fb5-4d4f-baa5-8b743e42255c" containerName="extract-utilities" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.477653 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="054f4be9-3fb5-4d4f-baa5-8b743e42255c" containerName="extract-utilities" Dec 11 08:52:42 crc kubenswrapper[4629]: E1211 08:52:42.477668 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="054f4be9-3fb5-4d4f-baa5-8b743e42255c" containerName="registry-server" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.477676 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="054f4be9-3fb5-4d4f-baa5-8b743e42255c" containerName="registry-server" Dec 11 08:52:42 crc kubenswrapper[4629]: E1211 08:52:42.477686 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77fc0b67-9596-4033-bc05-b0487f495258" containerName="extract-content" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.477691 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="77fc0b67-9596-4033-bc05-b0487f495258" containerName="extract-content" Dec 11 08:52:42 crc kubenswrapper[4629]: E1211 08:52:42.477699 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77fc0b67-9596-4033-bc05-b0487f495258" containerName="extract-utilities" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.477705 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="77fc0b67-9596-4033-bc05-b0487f495258" containerName="extract-utilities" Dec 11 08:52:42 crc kubenswrapper[4629]: E1211 08:52:42.477713 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="054f4be9-3fb5-4d4f-baa5-8b743e42255c" containerName="extract-content" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.477719 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="054f4be9-3fb5-4d4f-baa5-8b743e42255c" containerName="extract-content" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.477810 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="054f4be9-3fb5-4d4f-baa5-8b743e42255c" containerName="registry-server" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.477824 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="77fc0b67-9596-4033-bc05-b0487f495258" containerName="registry-server" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.478247 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2nxg4" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.480667 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-drbz8" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.482297 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.494823 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.501489 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2nxg4"] Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.585197 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbdqr\" (UniqueName: \"kubernetes.io/projected/03806eeb-2d32-41cd-8709-5aa61ae6c8c9-kube-api-access-hbdqr\") pod \"openstack-operator-index-2nxg4\" (UID: \"03806eeb-2d32-41cd-8709-5aa61ae6c8c9\") " pod="openstack-operators/openstack-operator-index-2nxg4" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.686290 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbdqr\" (UniqueName: \"kubernetes.io/projected/03806eeb-2d32-41cd-8709-5aa61ae6c8c9-kube-api-access-hbdqr\") pod \"openstack-operator-index-2nxg4\" (UID: \"03806eeb-2d32-41cd-8709-5aa61ae6c8c9\") " pod="openstack-operators/openstack-operator-index-2nxg4" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.711136 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbdqr\" (UniqueName: \"kubernetes.io/projected/03806eeb-2d32-41cd-8709-5aa61ae6c8c9-kube-api-access-hbdqr\") pod \"openstack-operator-index-2nxg4\" (UID: \"03806eeb-2d32-41cd-8709-5aa61ae6c8c9\") " pod="openstack-operators/openstack-operator-index-2nxg4" Dec 11 08:52:42 crc kubenswrapper[4629]: I1211 08:52:42.795736 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2nxg4" Dec 11 08:52:43 crc kubenswrapper[4629]: I1211 08:52:43.227965 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2nxg4"] Dec 11 08:52:43 crc kubenswrapper[4629]: I1211 08:52:43.873167 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2nxg4" event={"ID":"03806eeb-2d32-41cd-8709-5aa61ae6c8c9","Type":"ContainerStarted","Data":"bc6a22b28a9c3d36ef817e0d36b00e51a051741e2e61b99be173192ab0bd7507"} Dec 11 08:52:45 crc kubenswrapper[4629]: I1211 08:52:45.859765 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2nxg4"] Dec 11 08:52:45 crc kubenswrapper[4629]: I1211 08:52:45.888550 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2nxg4" event={"ID":"03806eeb-2d32-41cd-8709-5aa61ae6c8c9","Type":"ContainerStarted","Data":"d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd"} Dec 11 08:52:45 crc kubenswrapper[4629]: I1211 08:52:45.912463 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2nxg4" podStartSLOduration=1.5082823300000001 podStartE2EDuration="3.912446647s" podCreationTimestamp="2025-12-11 08:52:42 +0000 UTC" firstStartedPulling="2025-12-11 08:52:43.243698311 +0000 UTC m=+951.309113919" lastFinishedPulling="2025-12-11 08:52:45.647862628 +0000 UTC m=+953.713278236" observedRunningTime="2025-12-11 08:52:45.91094453 +0000 UTC m=+953.976360138" watchObservedRunningTime="2025-12-11 08:52:45.912446647 +0000 UTC m=+953.977862255" Dec 11 08:52:46 crc kubenswrapper[4629]: I1211 08:52:46.556200 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tf2sz" Dec 11 08:52:46 crc kubenswrapper[4629]: I1211 08:52:46.668611 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-kgkwp"] Dec 11 08:52:46 crc kubenswrapper[4629]: I1211 08:52:46.669748 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kgkwp" Dec 11 08:52:46 crc kubenswrapper[4629]: I1211 08:52:46.676861 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kgkwp"] Dec 11 08:52:46 crc kubenswrapper[4629]: I1211 08:52:46.738695 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd9b4\" (UniqueName: \"kubernetes.io/projected/1058bc3d-a814-4f48-b25f-d6750f55d459-kube-api-access-kd9b4\") pod \"openstack-operator-index-kgkwp\" (UID: \"1058bc3d-a814-4f48-b25f-d6750f55d459\") " pod="openstack-operators/openstack-operator-index-kgkwp" Dec 11 08:52:46 crc kubenswrapper[4629]: I1211 08:52:46.840106 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd9b4\" (UniqueName: \"kubernetes.io/projected/1058bc3d-a814-4f48-b25f-d6750f55d459-kube-api-access-kd9b4\") pod \"openstack-operator-index-kgkwp\" (UID: \"1058bc3d-a814-4f48-b25f-d6750f55d459\") " pod="openstack-operators/openstack-operator-index-kgkwp" Dec 11 08:52:46 crc kubenswrapper[4629]: I1211 08:52:46.878703 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd9b4\" (UniqueName: \"kubernetes.io/projected/1058bc3d-a814-4f48-b25f-d6750f55d459-kube-api-access-kd9b4\") pod \"openstack-operator-index-kgkwp\" (UID: \"1058bc3d-a814-4f48-b25f-d6750f55d459\") " pod="openstack-operators/openstack-operator-index-kgkwp" Dec 11 08:52:46 crc kubenswrapper[4629]: I1211 08:52:46.900280 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-2nxg4" podUID="03806eeb-2d32-41cd-8709-5aa61ae6c8c9" containerName="registry-server" containerID="cri-o://d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd" gracePeriod=2 Dec 11 08:52:46 crc kubenswrapper[4629]: I1211 08:52:46.987751 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kgkwp" Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.330569 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kgkwp"] Dec 11 08:52:47 crc kubenswrapper[4629]: W1211 08:52:47.342941 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1058bc3d_a814_4f48_b25f_d6750f55d459.slice/crio-4f2b747c1a632aa968b383bb0f464d82c12768b7ff8d0e24ffd045e2ca2ea356 WatchSource:0}: Error finding container 4f2b747c1a632aa968b383bb0f464d82c12768b7ff8d0e24ffd045e2ca2ea356: Status 404 returned error can't find the container with id 4f2b747c1a632aa968b383bb0f464d82c12768b7ff8d0e24ffd045e2ca2ea356 Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.428610 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2nxg4" Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.551630 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbdqr\" (UniqueName: \"kubernetes.io/projected/03806eeb-2d32-41cd-8709-5aa61ae6c8c9-kube-api-access-hbdqr\") pod \"03806eeb-2d32-41cd-8709-5aa61ae6c8c9\" (UID: \"03806eeb-2d32-41cd-8709-5aa61ae6c8c9\") " Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.557350 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03806eeb-2d32-41cd-8709-5aa61ae6c8c9-kube-api-access-hbdqr" (OuterVolumeSpecName: "kube-api-access-hbdqr") pod "03806eeb-2d32-41cd-8709-5aa61ae6c8c9" (UID: "03806eeb-2d32-41cd-8709-5aa61ae6c8c9"). InnerVolumeSpecName "kube-api-access-hbdqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.653973 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbdqr\" (UniqueName: \"kubernetes.io/projected/03806eeb-2d32-41cd-8709-5aa61ae6c8c9-kube-api-access-hbdqr\") on node \"crc\" DevicePath \"\"" Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.906792 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kgkwp" event={"ID":"1058bc3d-a814-4f48-b25f-d6750f55d459","Type":"ContainerStarted","Data":"76dd56aaab81a2d8bb482c694018fdbd4f2cac94d5f1de906e35775d77d9b7e7"} Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.907112 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kgkwp" event={"ID":"1058bc3d-a814-4f48-b25f-d6750f55d459","Type":"ContainerStarted","Data":"4f2b747c1a632aa968b383bb0f464d82c12768b7ff8d0e24ffd045e2ca2ea356"} Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.909596 4629 generic.go:334] "Generic (PLEG): container finished" podID="03806eeb-2d32-41cd-8709-5aa61ae6c8c9" containerID="d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd" exitCode=0 Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.909628 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2nxg4" event={"ID":"03806eeb-2d32-41cd-8709-5aa61ae6c8c9","Type":"ContainerDied","Data":"d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd"} Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.909645 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2nxg4" event={"ID":"03806eeb-2d32-41cd-8709-5aa61ae6c8c9","Type":"ContainerDied","Data":"bc6a22b28a9c3d36ef817e0d36b00e51a051741e2e61b99be173192ab0bd7507"} Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.909695 4629 scope.go:117] "RemoveContainer" containerID="d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd" Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.909777 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2nxg4" Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.939365 4629 scope.go:117] "RemoveContainer" containerID="d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd" Dec 11 08:52:47 crc kubenswrapper[4629]: E1211 08:52:47.939816 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd\": container with ID starting with d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd not found: ID does not exist" containerID="d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd" Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.939878 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd"} err="failed to get container status \"d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd\": rpc error: code = NotFound desc = could not find container \"d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd\": container with ID starting with d1cfeb77d25fc7bc8db83bf37f37efaac0ae3ae0694f0e90f21627c4a6f923bd not found: ID does not exist" Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.942471 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-kgkwp" podStartSLOduration=1.889690468 podStartE2EDuration="1.942453235s" podCreationTimestamp="2025-12-11 08:52:46 +0000 UTC" firstStartedPulling="2025-12-11 08:52:47.346987866 +0000 UTC m=+955.412403474" lastFinishedPulling="2025-12-11 08:52:47.399750633 +0000 UTC m=+955.465166241" observedRunningTime="2025-12-11 08:52:47.939501243 +0000 UTC m=+956.004916851" watchObservedRunningTime="2025-12-11 08:52:47.942453235 +0000 UTC m=+956.007868833" Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.955292 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2nxg4"] Dec 11 08:52:47 crc kubenswrapper[4629]: I1211 08:52:47.959025 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-2nxg4"] Dec 11 08:52:48 crc kubenswrapper[4629]: I1211 08:52:48.205248 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03806eeb-2d32-41cd-8709-5aa61ae6c8c9" path="/var/lib/kubelet/pods/03806eeb-2d32-41cd-8709-5aa61ae6c8c9/volumes" Dec 11 08:52:50 crc kubenswrapper[4629]: I1211 08:52:50.486213 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:52:50 crc kubenswrapper[4629]: I1211 08:52:50.486347 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:52:56 crc kubenswrapper[4629]: I1211 08:52:56.543101 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-t6bzz" Dec 11 08:52:56 crc kubenswrapper[4629]: I1211 08:52:56.988225 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-kgkwp" Dec 11 08:52:56 crc kubenswrapper[4629]: I1211 08:52:56.988615 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-kgkwp" Dec 11 08:52:57 crc kubenswrapper[4629]: I1211 08:52:57.022910 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-kgkwp" Dec 11 08:52:58 crc kubenswrapper[4629]: I1211 08:52:58.007344 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-kgkwp" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.498972 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v"] Dec 11 08:53:05 crc kubenswrapper[4629]: E1211 08:53:05.499720 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03806eeb-2d32-41cd-8709-5aa61ae6c8c9" containerName="registry-server" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.499732 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="03806eeb-2d32-41cd-8709-5aa61ae6c8c9" containerName="registry-server" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.499880 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="03806eeb-2d32-41cd-8709-5aa61ae6c8c9" containerName="registry-server" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.500755 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.503485 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-stbf7" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.515575 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v"] Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.648879 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6efefc09-794d-479a-b21a-b1c43506bcae-bundle\") pod \"3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v\" (UID: \"6efefc09-794d-479a-b21a-b1c43506bcae\") " pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.650003 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6efefc09-794d-479a-b21a-b1c43506bcae-util\") pod \"3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v\" (UID: \"6efefc09-794d-479a-b21a-b1c43506bcae\") " pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.650079 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwp4x\" (UniqueName: \"kubernetes.io/projected/6efefc09-794d-479a-b21a-b1c43506bcae-kube-api-access-vwp4x\") pod \"3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v\" (UID: \"6efefc09-794d-479a-b21a-b1c43506bcae\") " pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.751922 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6efefc09-794d-479a-b21a-b1c43506bcae-bundle\") pod \"3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v\" (UID: \"6efefc09-794d-479a-b21a-b1c43506bcae\") " pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.752288 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6efefc09-794d-479a-b21a-b1c43506bcae-util\") pod \"3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v\" (UID: \"6efefc09-794d-479a-b21a-b1c43506bcae\") " pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.752415 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwp4x\" (UniqueName: \"kubernetes.io/projected/6efefc09-794d-479a-b21a-b1c43506bcae-kube-api-access-vwp4x\") pod \"3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v\" (UID: \"6efefc09-794d-479a-b21a-b1c43506bcae\") " pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.752663 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6efefc09-794d-479a-b21a-b1c43506bcae-bundle\") pod \"3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v\" (UID: \"6efefc09-794d-479a-b21a-b1c43506bcae\") " pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.752709 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6efefc09-794d-479a-b21a-b1c43506bcae-util\") pod \"3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v\" (UID: \"6efefc09-794d-479a-b21a-b1c43506bcae\") " pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.771780 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwp4x\" (UniqueName: \"kubernetes.io/projected/6efefc09-794d-479a-b21a-b1c43506bcae-kube-api-access-vwp4x\") pod \"3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v\" (UID: \"6efefc09-794d-479a-b21a-b1c43506bcae\") " pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:05 crc kubenswrapper[4629]: I1211 08:53:05.830188 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:06 crc kubenswrapper[4629]: I1211 08:53:06.326030 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v"] Dec 11 08:53:07 crc kubenswrapper[4629]: I1211 08:53:07.039582 4629 generic.go:334] "Generic (PLEG): container finished" podID="6efefc09-794d-479a-b21a-b1c43506bcae" containerID="4427be425cbee3b12f36febdbfc3abd11008ed1494a855f2b5e49cdd3bbc666c" exitCode=0 Dec 11 08:53:07 crc kubenswrapper[4629]: I1211 08:53:07.039812 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" event={"ID":"6efefc09-794d-479a-b21a-b1c43506bcae","Type":"ContainerDied","Data":"4427be425cbee3b12f36febdbfc3abd11008ed1494a855f2b5e49cdd3bbc666c"} Dec 11 08:53:07 crc kubenswrapper[4629]: I1211 08:53:07.040145 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" event={"ID":"6efefc09-794d-479a-b21a-b1c43506bcae","Type":"ContainerStarted","Data":"9ef32b1a25cf82bc690a4d5467a9e687958f447e9b88321edf5f7f18c1334cf3"} Dec 11 08:53:08 crc kubenswrapper[4629]: I1211 08:53:08.048046 4629 generic.go:334] "Generic (PLEG): container finished" podID="6efefc09-794d-479a-b21a-b1c43506bcae" containerID="ee529e7508895615cacecb4e36de846927dda06c9d896973c008f07bb7c4d1ff" exitCode=0 Dec 11 08:53:08 crc kubenswrapper[4629]: I1211 08:53:08.048130 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" event={"ID":"6efefc09-794d-479a-b21a-b1c43506bcae","Type":"ContainerDied","Data":"ee529e7508895615cacecb4e36de846927dda06c9d896973c008f07bb7c4d1ff"} Dec 11 08:53:09 crc kubenswrapper[4629]: I1211 08:53:09.057235 4629 generic.go:334] "Generic (PLEG): container finished" podID="6efefc09-794d-479a-b21a-b1c43506bcae" containerID="fa8d112a07029df6d3441ba21b41ff72456b95557be15b01a6b9ddc3c0909e81" exitCode=0 Dec 11 08:53:09 crc kubenswrapper[4629]: I1211 08:53:09.057281 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" event={"ID":"6efefc09-794d-479a-b21a-b1c43506bcae","Type":"ContainerDied","Data":"fa8d112a07029df6d3441ba21b41ff72456b95557be15b01a6b9ddc3c0909e81"} Dec 11 08:53:10 crc kubenswrapper[4629]: I1211 08:53:10.292042 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:10 crc kubenswrapper[4629]: I1211 08:53:10.422701 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6efefc09-794d-479a-b21a-b1c43506bcae-bundle\") pod \"6efefc09-794d-479a-b21a-b1c43506bcae\" (UID: \"6efefc09-794d-479a-b21a-b1c43506bcae\") " Dec 11 08:53:10 crc kubenswrapper[4629]: I1211 08:53:10.422897 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwp4x\" (UniqueName: \"kubernetes.io/projected/6efefc09-794d-479a-b21a-b1c43506bcae-kube-api-access-vwp4x\") pod \"6efefc09-794d-479a-b21a-b1c43506bcae\" (UID: \"6efefc09-794d-479a-b21a-b1c43506bcae\") " Dec 11 08:53:10 crc kubenswrapper[4629]: I1211 08:53:10.422920 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6efefc09-794d-479a-b21a-b1c43506bcae-util\") pod \"6efefc09-794d-479a-b21a-b1c43506bcae\" (UID: \"6efefc09-794d-479a-b21a-b1c43506bcae\") " Dec 11 08:53:10 crc kubenswrapper[4629]: I1211 08:53:10.423744 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6efefc09-794d-479a-b21a-b1c43506bcae-bundle" (OuterVolumeSpecName: "bundle") pod "6efefc09-794d-479a-b21a-b1c43506bcae" (UID: "6efefc09-794d-479a-b21a-b1c43506bcae"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:53:10 crc kubenswrapper[4629]: I1211 08:53:10.433119 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6efefc09-794d-479a-b21a-b1c43506bcae-kube-api-access-vwp4x" (OuterVolumeSpecName: "kube-api-access-vwp4x") pod "6efefc09-794d-479a-b21a-b1c43506bcae" (UID: "6efefc09-794d-479a-b21a-b1c43506bcae"). InnerVolumeSpecName "kube-api-access-vwp4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:53:10 crc kubenswrapper[4629]: I1211 08:53:10.436598 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6efefc09-794d-479a-b21a-b1c43506bcae-util" (OuterVolumeSpecName: "util") pod "6efefc09-794d-479a-b21a-b1c43506bcae" (UID: "6efefc09-794d-479a-b21a-b1c43506bcae"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:53:10 crc kubenswrapper[4629]: I1211 08:53:10.524068 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwp4x\" (UniqueName: \"kubernetes.io/projected/6efefc09-794d-479a-b21a-b1c43506bcae-kube-api-access-vwp4x\") on node \"crc\" DevicePath \"\"" Dec 11 08:53:10 crc kubenswrapper[4629]: I1211 08:53:10.524135 4629 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6efefc09-794d-479a-b21a-b1c43506bcae-util\") on node \"crc\" DevicePath \"\"" Dec 11 08:53:10 crc kubenswrapper[4629]: I1211 08:53:10.524145 4629 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6efefc09-794d-479a-b21a-b1c43506bcae-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:53:11 crc kubenswrapper[4629]: I1211 08:53:11.070143 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" event={"ID":"6efefc09-794d-479a-b21a-b1c43506bcae","Type":"ContainerDied","Data":"9ef32b1a25cf82bc690a4d5467a9e687958f447e9b88321edf5f7f18c1334cf3"} Dec 11 08:53:11 crc kubenswrapper[4629]: I1211 08:53:11.070488 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ef32b1a25cf82bc690a4d5467a9e687958f447e9b88321edf5f7f18c1334cf3" Dec 11 08:53:11 crc kubenswrapper[4629]: I1211 08:53:11.070230 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v" Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.058117 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk"] Dec 11 08:53:18 crc kubenswrapper[4629]: E1211 08:53:18.060268 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6efefc09-794d-479a-b21a-b1c43506bcae" containerName="extract" Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.060380 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="6efefc09-794d-479a-b21a-b1c43506bcae" containerName="extract" Dec 11 08:53:18 crc kubenswrapper[4629]: E1211 08:53:18.060495 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6efefc09-794d-479a-b21a-b1c43506bcae" containerName="pull" Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.060589 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="6efefc09-794d-479a-b21a-b1c43506bcae" containerName="pull" Dec 11 08:53:18 crc kubenswrapper[4629]: E1211 08:53:18.060676 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6efefc09-794d-479a-b21a-b1c43506bcae" containerName="util" Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.060757 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="6efefc09-794d-479a-b21a-b1c43506bcae" containerName="util" Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.061012 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="6efefc09-794d-479a-b21a-b1c43506bcae" containerName="extract" Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.061632 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk" Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.065118 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-r22zk" Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.078138 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk"] Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.126047 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w7q6\" (UniqueName: \"kubernetes.io/projected/9dc7763f-4b47-4337-b320-fc46f74c2ed3-kube-api-access-6w7q6\") pod \"openstack-operator-controller-operator-7b8547f8bd-jz7pk\" (UID: \"9dc7763f-4b47-4337-b320-fc46f74c2ed3\") " pod="openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk" Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.227801 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w7q6\" (UniqueName: \"kubernetes.io/projected/9dc7763f-4b47-4337-b320-fc46f74c2ed3-kube-api-access-6w7q6\") pod \"openstack-operator-controller-operator-7b8547f8bd-jz7pk\" (UID: \"9dc7763f-4b47-4337-b320-fc46f74c2ed3\") " pod="openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk" Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.249822 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w7q6\" (UniqueName: \"kubernetes.io/projected/9dc7763f-4b47-4337-b320-fc46f74c2ed3-kube-api-access-6w7q6\") pod \"openstack-operator-controller-operator-7b8547f8bd-jz7pk\" (UID: \"9dc7763f-4b47-4337-b320-fc46f74c2ed3\") " pod="openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk" Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.389489 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk" Dec 11 08:53:18 crc kubenswrapper[4629]: I1211 08:53:18.837905 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk"] Dec 11 08:53:19 crc kubenswrapper[4629]: I1211 08:53:19.120223 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk" event={"ID":"9dc7763f-4b47-4337-b320-fc46f74c2ed3","Type":"ContainerStarted","Data":"9b577dfd073501d7778910e5a37ec2784e89e82a9bfec9cb046a86336f124301"} Dec 11 08:53:20 crc kubenswrapper[4629]: I1211 08:53:20.486200 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:53:20 crc kubenswrapper[4629]: I1211 08:53:20.486581 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:53:20 crc kubenswrapper[4629]: I1211 08:53:20.486654 4629 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:53:20 crc kubenswrapper[4629]: I1211 08:53:20.487396 4629 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3478d1305c4015a5af7c390010bcd373d9a5066e2359bd5ed303da8f30af223b"} pod="openshift-machine-config-operator/machine-config-daemon-wbld9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:53:20 crc kubenswrapper[4629]: I1211 08:53:20.487465 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" containerID="cri-o://3478d1305c4015a5af7c390010bcd373d9a5066e2359bd5ed303da8f30af223b" gracePeriod=600 Dec 11 08:53:21 crc kubenswrapper[4629]: I1211 08:53:21.143435 4629 generic.go:334] "Generic (PLEG): container finished" podID="847b50b9-0093-404f-b0dd-63c814d8986b" containerID="3478d1305c4015a5af7c390010bcd373d9a5066e2359bd5ed303da8f30af223b" exitCode=0 Dec 11 08:53:21 crc kubenswrapper[4629]: I1211 08:53:21.143486 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerDied","Data":"3478d1305c4015a5af7c390010bcd373d9a5066e2359bd5ed303da8f30af223b"} Dec 11 08:53:21 crc kubenswrapper[4629]: I1211 08:53:21.143525 4629 scope.go:117] "RemoveContainer" containerID="3a42a64840fd258d581b9d590c1921c344fcce2923f32019baf2139ce7bfec2f" Dec 11 08:53:27 crc kubenswrapper[4629]: I1211 08:53:27.214557 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"921d43c980dda9b89debb49d73c6ac3dda5f931588d36bb68efdef51a59b46b8"} Dec 11 08:53:27 crc kubenswrapper[4629]: I1211 08:53:27.221293 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk" event={"ID":"9dc7763f-4b47-4337-b320-fc46f74c2ed3","Type":"ContainerStarted","Data":"e8650d0bcfdf215e75068b94f272dd132a827c882bae0666c095b3808559140d"} Dec 11 08:53:27 crc kubenswrapper[4629]: I1211 08:53:27.221541 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk" Dec 11 08:53:27 crc kubenswrapper[4629]: I1211 08:53:27.269426 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk" podStartSLOduration=1.884505577 podStartE2EDuration="9.269387713s" podCreationTimestamp="2025-12-11 08:53:18 +0000 UTC" firstStartedPulling="2025-12-11 08:53:18.846312454 +0000 UTC m=+986.911728062" lastFinishedPulling="2025-12-11 08:53:26.23119459 +0000 UTC m=+994.296610198" observedRunningTime="2025-12-11 08:53:27.265515862 +0000 UTC m=+995.330931480" watchObservedRunningTime="2025-12-11 08:53:27.269387713 +0000 UTC m=+995.334803331" Dec 11 08:53:38 crc kubenswrapper[4629]: I1211 08:53:38.392493 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7b8547f8bd-jz7pk" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.653273 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.655010 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.659262 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-q2tgm" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.666762 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.668790 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.672512 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-g7cxp" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.672816 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.679535 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.680630 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.686441 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-k7qrn" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.688357 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhv7n\" (UniqueName: \"kubernetes.io/projected/39798d73-c400-4797-b5ac-e402d17a1df1-kube-api-access-zhv7n\") pod \"cinder-operator-controller-manager-6c677c69b-p8lh9\" (UID: \"39798d73-c400-4797-b5ac-e402d17a1df1\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.688470 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwh6d\" (UniqueName: \"kubernetes.io/projected/813d511c-6f11-40c8-a9c3-281e4909c88d-kube-api-access-cwh6d\") pod \"barbican-operator-controller-manager-7d9dfd778-g5zpd\" (UID: \"813d511c-6f11-40c8-a9c3-281e4909c88d\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.691358 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.725360 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.745679 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.771313 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.780733 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-ftvvf" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.807977 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nsxg\" (UniqueName: \"kubernetes.io/projected/64242f6c-6d25-4868-acd8-de067ecdeb7a-kube-api-access-9nsxg\") pod \"glance-operator-controller-manager-5697bb5779-xdp6j\" (UID: \"64242f6c-6d25-4868-acd8-de067ecdeb7a\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.808063 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhv7n\" (UniqueName: \"kubernetes.io/projected/39798d73-c400-4797-b5ac-e402d17a1df1-kube-api-access-zhv7n\") pod \"cinder-operator-controller-manager-6c677c69b-p8lh9\" (UID: \"39798d73-c400-4797-b5ac-e402d17a1df1\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.808169 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwh6d\" (UniqueName: \"kubernetes.io/projected/813d511c-6f11-40c8-a9c3-281e4909c88d-kube-api-access-cwh6d\") pod \"barbican-operator-controller-manager-7d9dfd778-g5zpd\" (UID: \"813d511c-6f11-40c8-a9c3-281e4909c88d\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.808229 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl9nj\" (UniqueName: \"kubernetes.io/projected/c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1-kube-api-access-hl9nj\") pod \"designate-operator-controller-manager-697fb699cf-mhqnx\" (UID: \"c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.813661 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.830226 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.831741 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.844290 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-zl6zj" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.854557 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.877336 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwh6d\" (UniqueName: \"kubernetes.io/projected/813d511c-6f11-40c8-a9c3-281e4909c88d-kube-api-access-cwh6d\") pod \"barbican-operator-controller-manager-7d9dfd778-g5zpd\" (UID: \"813d511c-6f11-40c8-a9c3-281e4909c88d\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.877583 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhv7n\" (UniqueName: \"kubernetes.io/projected/39798d73-c400-4797-b5ac-e402d17a1df1-kube-api-access-zhv7n\") pod \"cinder-operator-controller-manager-6c677c69b-p8lh9\" (UID: \"39798d73-c400-4797-b5ac-e402d17a1df1\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.911479 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvxh8\" (UniqueName: \"kubernetes.io/projected/8a37b92a-08fc-47a9-8d55-97182286daea-kube-api-access-cvxh8\") pod \"heat-operator-controller-manager-5f64f6f8bb-8mtlt\" (UID: \"8a37b92a-08fc-47a9-8d55-97182286daea\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.911564 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl9nj\" (UniqueName: \"kubernetes.io/projected/c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1-kube-api-access-hl9nj\") pod \"designate-operator-controller-manager-697fb699cf-mhqnx\" (UID: \"c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.911594 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nsxg\" (UniqueName: \"kubernetes.io/projected/64242f6c-6d25-4868-acd8-de067ecdeb7a-kube-api-access-9nsxg\") pod \"glance-operator-controller-manager-5697bb5779-xdp6j\" (UID: \"64242f6c-6d25-4868-acd8-de067ecdeb7a\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.922819 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.924282 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.952185 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.970642 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-rgxhm" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.981375 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nsxg\" (UniqueName: \"kubernetes.io/projected/64242f6c-6d25-4868-acd8-de067ecdeb7a-kube-api-access-9nsxg\") pod \"glance-operator-controller-manager-5697bb5779-xdp6j\" (UID: \"64242f6c-6d25-4868-acd8-de067ecdeb7a\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.981663 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.983172 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.983611 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.985186 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-6lqs7" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.985609 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.986409 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw"] Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.987412 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.994755 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl9nj\" (UniqueName: \"kubernetes.io/projected/c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1-kube-api-access-hl9nj\") pod \"designate-operator-controller-manager-697fb699cf-mhqnx\" (UID: \"c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx" Dec 11 08:54:06 crc kubenswrapper[4629]: I1211 08:54:06.996217 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.000126 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.000789 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.001260 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-bkhl2" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.007667 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.012920 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-526gc\" (UniqueName: \"kubernetes.io/projected/219ddd48-017d-4da4-b576-0b8f24bc40be-kube-api-access-526gc\") pod \"infra-operator-controller-manager-78d48bff9d-l59gh\" (UID: \"219ddd48-017d-4da4-b576-0b8f24bc40be\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.012984 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjcfc\" (UniqueName: \"kubernetes.io/projected/7833d1b6-8dbd-4afa-9042-755b5a9c383b-kube-api-access-jjcfc\") pod \"horizon-operator-controller-manager-68c6d99b8f-kq7vj\" (UID: \"7833d1b6-8dbd-4afa-9042-755b5a9c383b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.013024 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert\") pod \"infra-operator-controller-manager-78d48bff9d-l59gh\" (UID: \"219ddd48-017d-4da4-b576-0b8f24bc40be\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.013061 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvxh8\" (UniqueName: \"kubernetes.io/projected/8a37b92a-08fc-47a9-8d55-97182286daea-kube-api-access-cvxh8\") pod \"heat-operator-controller-manager-5f64f6f8bb-8mtlt\" (UID: \"8a37b92a-08fc-47a9-8d55-97182286daea\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.038811 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.056301 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.057096 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-q9pzz" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.079934 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.097324 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.099639 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.108804 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-4pwh2" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.116250 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-526gc\" (UniqueName: \"kubernetes.io/projected/219ddd48-017d-4da4-b576-0b8f24bc40be-kube-api-access-526gc\") pod \"infra-operator-controller-manager-78d48bff9d-l59gh\" (UID: \"219ddd48-017d-4da4-b576-0b8f24bc40be\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.117420 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbcvx\" (UniqueName: \"kubernetes.io/projected/94ec218d-3b08-463d-8225-f12cbcf0eb8d-kube-api-access-mbcvx\") pod \"keystone-operator-controller-manager-7765d96ddf-kqcvw\" (UID: \"94ec218d-3b08-463d-8225-f12cbcf0eb8d\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.133664 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvxh8\" (UniqueName: \"kubernetes.io/projected/8a37b92a-08fc-47a9-8d55-97182286daea-kube-api-access-cvxh8\") pod \"heat-operator-controller-manager-5f64f6f8bb-8mtlt\" (UID: \"8a37b92a-08fc-47a9-8d55-97182286daea\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.133974 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjcfc\" (UniqueName: \"kubernetes.io/projected/7833d1b6-8dbd-4afa-9042-755b5a9c383b-kube-api-access-jjcfc\") pod \"horizon-operator-controller-manager-68c6d99b8f-kq7vj\" (UID: \"7833d1b6-8dbd-4afa-9042-755b5a9c383b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.134540 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert\") pod \"infra-operator-controller-manager-78d48bff9d-l59gh\" (UID: \"219ddd48-017d-4da4-b576-0b8f24bc40be\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.134676 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mz5m\" (UniqueName: \"kubernetes.io/projected/bf19951f-862f-4b74-92ab-63cdede5be6d-kube-api-access-8mz5m\") pod \"ironic-operator-controller-manager-967d97867-qxsl5\" (UID: \"bf19951f-862f-4b74-92ab-63cdede5be6d\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.130776 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j" Dec 11 08:54:07 crc kubenswrapper[4629]: E1211 08:54:07.135087 4629 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 08:54:07 crc kubenswrapper[4629]: E1211 08:54:07.135220 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert podName:219ddd48-017d-4da4-b576-0b8f24bc40be nodeName:}" failed. No retries permitted until 2025-12-11 08:54:07.635178843 +0000 UTC m=+1035.700594451 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert") pod "infra-operator-controller-manager-78d48bff9d-l59gh" (UID: "219ddd48-017d-4da4-b576-0b8f24bc40be") : secret "infra-operator-webhook-server-cert" not found Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.147931 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.149385 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.164437 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qpgjx" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.169203 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.180816 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-526gc\" (UniqueName: \"kubernetes.io/projected/219ddd48-017d-4da4-b576-0b8f24bc40be-kube-api-access-526gc\") pod \"infra-operator-controller-manager-78d48bff9d-l59gh\" (UID: \"219ddd48-017d-4da4-b576-0b8f24bc40be\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.187705 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjcfc\" (UniqueName: \"kubernetes.io/projected/7833d1b6-8dbd-4afa-9042-755b5a9c383b-kube-api-access-jjcfc\") pod \"horizon-operator-controller-manager-68c6d99b8f-kq7vj\" (UID: \"7833d1b6-8dbd-4afa-9042-755b5a9c383b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.202705 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.216522 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.217699 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.233335 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-xn7tx" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.239797 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mz5m\" (UniqueName: \"kubernetes.io/projected/bf19951f-862f-4b74-92ab-63cdede5be6d-kube-api-access-8mz5m\") pod \"ironic-operator-controller-manager-967d97867-qxsl5\" (UID: \"bf19951f-862f-4b74-92ab-63cdede5be6d\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.240157 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg495\" (UniqueName: \"kubernetes.io/projected/0f40457d-36a2-4183-8434-9cb4688489bb-kube-api-access-cg495\") pod \"manila-operator-controller-manager-5b5fd79c9c-9m9lh\" (UID: \"0f40457d-36a2-4183-8434-9cb4688489bb\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.240290 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbcvx\" (UniqueName: \"kubernetes.io/projected/94ec218d-3b08-463d-8225-f12cbcf0eb8d-kube-api-access-mbcvx\") pod \"keystone-operator-controller-manager-7765d96ddf-kqcvw\" (UID: \"94ec218d-3b08-463d-8225-f12cbcf0eb8d\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.240457 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mxjt\" (UniqueName: \"kubernetes.io/projected/bf9c412a-98bd-4152-b65a-637822c14b67-kube-api-access-2mxjt\") pod \"mariadb-operator-controller-manager-79c8c4686c-rmlvk\" (UID: \"bf9c412a-98bd-4152-b65a-637822c14b67\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.276275 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.296050 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.324413 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.332923 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.334499 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.348149 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.348945 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg495\" (UniqueName: \"kubernetes.io/projected/0f40457d-36a2-4183-8434-9cb4688489bb-kube-api-access-cg495\") pod \"manila-operator-controller-manager-5b5fd79c9c-9m9lh\" (UID: \"0f40457d-36a2-4183-8434-9cb4688489bb\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.349015 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fk84\" (UniqueName: \"kubernetes.io/projected/8c62910d-1b99-40df-be12-c3eb86c645aa-kube-api-access-7fk84\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-mjhzk\" (UID: \"8c62910d-1b99-40df-be12-c3eb86c645aa\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.349051 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mxjt\" (UniqueName: \"kubernetes.io/projected/bf9c412a-98bd-4152-b65a-637822c14b67-kube-api-access-2mxjt\") pod \"mariadb-operator-controller-manager-79c8c4686c-rmlvk\" (UID: \"bf9c412a-98bd-4152-b65a-637822c14b67\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.351542 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.353463 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.378948 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.399454 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.401895 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.404564 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.420992 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.424727 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.425043 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-x7x4q" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.425226 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-z5gs9" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.425494 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-8j52m" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.448096 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.464686 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fr7wqh\" (UID: \"49030eaa-0443-4db5-bfd2-3c4e17a3780e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.464764 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdl8d\" (UniqueName: \"kubernetes.io/projected/49030eaa-0443-4db5-bfd2-3c4e17a3780e-kube-api-access-hdl8d\") pod \"openstack-baremetal-operator-controller-manager-84b575879fr7wqh\" (UID: \"49030eaa-0443-4db5-bfd2-3c4e17a3780e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.464805 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fk84\" (UniqueName: \"kubernetes.io/projected/8c62910d-1b99-40df-be12-c3eb86c645aa-kube-api-access-7fk84\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-mjhzk\" (UID: \"8c62910d-1b99-40df-be12-c3eb86c645aa\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.464879 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zsj8\" (UniqueName: \"kubernetes.io/projected/99ac9552-d81c-47b9-a77c-c18fae357872-kube-api-access-6zsj8\") pod \"nova-operator-controller-manager-697bc559fc-p9wjh\" (UID: \"99ac9552-d81c-47b9-a77c-c18fae357872\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.464928 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9j6r\" (UniqueName: \"kubernetes.io/projected/3daa4e0a-c804-4d70-a79c-5e4d5dd3a602-kube-api-access-f9j6r\") pod \"octavia-operator-controller-manager-998648c74-tg2w6\" (UID: \"3daa4e0a-c804-4d70-a79c-5e4d5dd3a602\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.466564 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-l8jnf" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.485147 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mz5m\" (UniqueName: \"kubernetes.io/projected/bf19951f-862f-4b74-92ab-63cdede5be6d-kube-api-access-8mz5m\") pod \"ironic-operator-controller-manager-967d97867-qxsl5\" (UID: \"bf19951f-862f-4b74-92ab-63cdede5be6d\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.488248 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mxjt\" (UniqueName: \"kubernetes.io/projected/bf9c412a-98bd-4152-b65a-637822c14b67-kube-api-access-2mxjt\") pod \"mariadb-operator-controller-manager-79c8c4686c-rmlvk\" (UID: \"bf9c412a-98bd-4152-b65a-637822c14b67\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.533688 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg495\" (UniqueName: \"kubernetes.io/projected/0f40457d-36a2-4183-8434-9cb4688489bb-kube-api-access-cg495\") pod \"manila-operator-controller-manager-5b5fd79c9c-9m9lh\" (UID: \"0f40457d-36a2-4183-8434-9cb4688489bb\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.534736 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.534786 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbcvx\" (UniqueName: \"kubernetes.io/projected/94ec218d-3b08-463d-8225-f12cbcf0eb8d-kube-api-access-mbcvx\") pod \"keystone-operator-controller-manager-7765d96ddf-kqcvw\" (UID: \"94ec218d-3b08-463d-8225-f12cbcf0eb8d\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.569472 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fr7wqh\" (UID: \"49030eaa-0443-4db5-bfd2-3c4e17a3780e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.569561 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdl8d\" (UniqueName: \"kubernetes.io/projected/49030eaa-0443-4db5-bfd2-3c4e17a3780e-kube-api-access-hdl8d\") pod \"openstack-baremetal-operator-controller-manager-84b575879fr7wqh\" (UID: \"49030eaa-0443-4db5-bfd2-3c4e17a3780e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.569694 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zsj8\" (UniqueName: \"kubernetes.io/projected/99ac9552-d81c-47b9-a77c-c18fae357872-kube-api-access-6zsj8\") pod \"nova-operator-controller-manager-697bc559fc-p9wjh\" (UID: \"99ac9552-d81c-47b9-a77c-c18fae357872\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.569737 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj4lm\" (UniqueName: \"kubernetes.io/projected/c808480f-8564-4628-b9b9-7ee9700cbb91-kube-api-access-tj4lm\") pod \"ovn-operator-controller-manager-b6456fdb6-f5ftx\" (UID: \"c808480f-8564-4628-b9b9-7ee9700cbb91\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.569799 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9j6r\" (UniqueName: \"kubernetes.io/projected/3daa4e0a-c804-4d70-a79c-5e4d5dd3a602-kube-api-access-f9j6r\") pod \"octavia-operator-controller-manager-998648c74-tg2w6\" (UID: \"3daa4e0a-c804-4d70-a79c-5e4d5dd3a602\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" Dec 11 08:54:07 crc kubenswrapper[4629]: E1211 08:54:07.570322 4629 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:54:07 crc kubenswrapper[4629]: E1211 08:54:07.570398 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert podName:49030eaa-0443-4db5-bfd2-3c4e17a3780e nodeName:}" failed. No retries permitted until 2025-12-11 08:54:08.07036808 +0000 UTC m=+1036.135783688 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fr7wqh" (UID: "49030eaa-0443-4db5-bfd2-3c4e17a3780e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.704816 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj4lm\" (UniqueName: \"kubernetes.io/projected/c808480f-8564-4628-b9b9-7ee9700cbb91-kube-api-access-tj4lm\") pod \"ovn-operator-controller-manager-b6456fdb6-f5ftx\" (UID: \"c808480f-8564-4628-b9b9-7ee9700cbb91\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.706053 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert\") pod \"infra-operator-controller-manager-78d48bff9d-l59gh\" (UID: \"219ddd48-017d-4da4-b576-0b8f24bc40be\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.706286 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zsj8\" (UniqueName: \"kubernetes.io/projected/99ac9552-d81c-47b9-a77c-c18fae357872-kube-api-access-6zsj8\") pod \"nova-operator-controller-manager-697bc559fc-p9wjh\" (UID: \"99ac9552-d81c-47b9-a77c-c18fae357872\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh" Dec 11 08:54:07 crc kubenswrapper[4629]: E1211 08:54:07.710771 4629 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 08:54:07 crc kubenswrapper[4629]: E1211 08:54:07.710961 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert podName:219ddd48-017d-4da4-b576-0b8f24bc40be nodeName:}" failed. No retries permitted until 2025-12-11 08:54:08.710882233 +0000 UTC m=+1036.776297841 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert") pod "infra-operator-controller-manager-78d48bff9d-l59gh" (UID: "219ddd48-017d-4da4-b576-0b8f24bc40be") : secret "infra-operator-webhook-server-cert" not found Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.711723 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.716082 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fk84\" (UniqueName: \"kubernetes.io/projected/8c62910d-1b99-40df-be12-c3eb86c645aa-kube-api-access-7fk84\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-mjhzk\" (UID: \"8c62910d-1b99-40df-be12-c3eb86c645aa\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.765874 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9j6r\" (UniqueName: \"kubernetes.io/projected/3daa4e0a-c804-4d70-a79c-5e4d5dd3a602-kube-api-access-f9j6r\") pod \"octavia-operator-controller-manager-998648c74-tg2w6\" (UID: \"3daa4e0a-c804-4d70-a79c-5e4d5dd3a602\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.768760 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj4lm\" (UniqueName: \"kubernetes.io/projected/c808480f-8564-4628-b9b9-7ee9700cbb91-kube-api-access-tj4lm\") pod \"ovn-operator-controller-manager-b6456fdb6-f5ftx\" (UID: \"c808480f-8564-4628-b9b9-7ee9700cbb91\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.784954 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.789705 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.800074 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.835134 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.925716 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdl8d\" (UniqueName: \"kubernetes.io/projected/49030eaa-0443-4db5-bfd2-3c4e17a3780e-kube-api-access-hdl8d\") pod \"openstack-baremetal-operator-controller-manager-84b575879fr7wqh\" (UID: \"49030eaa-0443-4db5-bfd2-3c4e17a3780e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.928390 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.930252 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.931732 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.932524 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj4fq\" (UniqueName: \"kubernetes.io/projected/8ebecdcb-ee18-4285-8b09-65c8859dc77e-kube-api-access-tj4fq\") pod \"swift-operator-controller-manager-9d58d64bc-gtqck\" (UID: \"8ebecdcb-ee18-4285-8b09-65c8859dc77e\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.932883 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.934569 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.947497 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-jzqjn" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.947717 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.949837 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.962215 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-2c8cn" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.963890 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.980518 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh"] Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.982305 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" Dec 11 08:54:07 crc kubenswrapper[4629]: I1211 08:54:07.993051 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-9v2jz" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.016303 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk"] Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.036946 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj4fq\" (UniqueName: \"kubernetes.io/projected/8ebecdcb-ee18-4285-8b09-65c8859dc77e-kube-api-access-tj4fq\") pod \"swift-operator-controller-manager-9d58d64bc-gtqck\" (UID: \"8ebecdcb-ee18-4285-8b09-65c8859dc77e\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.045192 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh"] Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.066029 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-f8m45"] Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.067497 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-f8m45" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.073799 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-p6gvb" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.096230 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-f8m45"] Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.147504 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2llp\" (UniqueName: \"kubernetes.io/projected/2382ec97-3dac-4285-bb95-fa1864b1bd8d-kube-api-access-f2llp\") pod \"test-operator-controller-manager-5854674fcc-f8m45\" (UID: \"2382ec97-3dac-4285-bb95-fa1864b1bd8d\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-f8m45" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.147921 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xs8x\" (UniqueName: \"kubernetes.io/projected/8e2cace8-e27c-4d40-b20e-dd16cb10a425-kube-api-access-5xs8x\") pod \"placement-operator-controller-manager-78f8948974-8q2kk\" (UID: \"8e2cace8-e27c-4d40-b20e-dd16cb10a425\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.148156 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6fp2\" (UniqueName: \"kubernetes.io/projected/1150e5d5-ae4c-4800-a10d-9ba271074338-kube-api-access-v6fp2\") pod \"telemetry-operator-controller-manager-58d5ff84df-cptvh\" (UID: \"1150e5d5-ae4c-4800-a10d-9ba271074338\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.148461 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fr7wqh\" (UID: \"49030eaa-0443-4db5-bfd2-3c4e17a3780e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:08 crc kubenswrapper[4629]: E1211 08:54:08.148880 4629 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:54:08 crc kubenswrapper[4629]: E1211 08:54:08.149093 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert podName:49030eaa-0443-4db5-bfd2-3c4e17a3780e nodeName:}" failed. No retries permitted until 2025-12-11 08:54:09.149059852 +0000 UTC m=+1037.214475460 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fr7wqh" (UID: "49030eaa-0443-4db5-bfd2-3c4e17a3780e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.369840 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6fp2\" (UniqueName: \"kubernetes.io/projected/1150e5d5-ae4c-4800-a10d-9ba271074338-kube-api-access-v6fp2\") pod \"telemetry-operator-controller-manager-58d5ff84df-cptvh\" (UID: \"1150e5d5-ae4c-4800-a10d-9ba271074338\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.370274 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2llp\" (UniqueName: \"kubernetes.io/projected/2382ec97-3dac-4285-bb95-fa1864b1bd8d-kube-api-access-f2llp\") pod \"test-operator-controller-manager-5854674fcc-f8m45\" (UID: \"2382ec97-3dac-4285-bb95-fa1864b1bd8d\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-f8m45" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.370398 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xs8x\" (UniqueName: \"kubernetes.io/projected/8e2cace8-e27c-4d40-b20e-dd16cb10a425-kube-api-access-5xs8x\") pod \"placement-operator-controller-manager-78f8948974-8q2kk\" (UID: \"8e2cace8-e27c-4d40-b20e-dd16cb10a425\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.409609 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72"] Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.411006 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72"] Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.411096 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.424023 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj4fq\" (UniqueName: \"kubernetes.io/projected/8ebecdcb-ee18-4285-8b09-65c8859dc77e-kube-api-access-tj4fq\") pod \"swift-operator-controller-manager-9d58d64bc-gtqck\" (UID: \"8ebecdcb-ee18-4285-8b09-65c8859dc77e\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.466958 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-mkkqw" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.550942 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6fp2\" (UniqueName: \"kubernetes.io/projected/1150e5d5-ae4c-4800-a10d-9ba271074338-kube-api-access-v6fp2\") pod \"telemetry-operator-controller-manager-58d5ff84df-cptvh\" (UID: \"1150e5d5-ae4c-4800-a10d-9ba271074338\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.556727 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2llp\" (UniqueName: \"kubernetes.io/projected/2382ec97-3dac-4285-bb95-fa1864b1bd8d-kube-api-access-f2llp\") pod \"test-operator-controller-manager-5854674fcc-f8m45\" (UID: \"2382ec97-3dac-4285-bb95-fa1864b1bd8d\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-f8m45" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.573256 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n8nm\" (UniqueName: \"kubernetes.io/projected/a6ef2a86-805b-44ed-95bd-1c53e457b9d9-kube-api-access-9n8nm\") pod \"watcher-operator-controller-manager-75944c9b7-8qz72\" (UID: \"a6ef2a86-805b-44ed-95bd-1c53e457b9d9\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.578455 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xs8x\" (UniqueName: \"kubernetes.io/projected/8e2cace8-e27c-4d40-b20e-dd16cb10a425-kube-api-access-5xs8x\") pod \"placement-operator-controller-manager-78f8948974-8q2kk\" (UID: \"8e2cace8-e27c-4d40-b20e-dd16cb10a425\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.606098 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.839121 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n8nm\" (UniqueName: \"kubernetes.io/projected/a6ef2a86-805b-44ed-95bd-1c53e457b9d9-kube-api-access-9n8nm\") pod \"watcher-operator-controller-manager-75944c9b7-8qz72\" (UID: \"a6ef2a86-805b-44ed-95bd-1c53e457b9d9\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.839269 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert\") pod \"infra-operator-controller-manager-78d48bff9d-l59gh\" (UID: \"219ddd48-017d-4da4-b576-0b8f24bc40be\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:08 crc kubenswrapper[4629]: E1211 08:54:08.839400 4629 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 08:54:08 crc kubenswrapper[4629]: E1211 08:54:08.839455 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert podName:219ddd48-017d-4da4-b576-0b8f24bc40be nodeName:}" failed. No retries permitted until 2025-12-11 08:54:10.839438301 +0000 UTC m=+1038.904853909 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert") pod "infra-operator-controller-manager-78d48bff9d-l59gh" (UID: "219ddd48-017d-4da4-b576-0b8f24bc40be") : secret "infra-operator-webhook-server-cert" not found Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.849087 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.927716 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b"] Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.928337 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.942885 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b"] Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.943009 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.949547 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.950276 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n8nm\" (UniqueName: \"kubernetes.io/projected/a6ef2a86-805b-44ed-95bd-1c53e457b9d9-kube-api-access-9n8nm\") pod \"watcher-operator-controller-manager-75944c9b7-8qz72\" (UID: \"a6ef2a86-805b-44ed-95bd-1c53e457b9d9\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.950340 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-jt77w" Dec 11 08:54:08 crc kubenswrapper[4629]: I1211 08:54:08.950384 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.202542 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.202581 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.202601 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2f78\" (UniqueName: \"kubernetes.io/projected/914206a5-adb8-4482-90f6-e55a294259c3-kube-api-access-c2f78\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.202648 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fr7wqh\" (UID: \"49030eaa-0443-4db5-bfd2-3c4e17a3780e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:09.202832 4629 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:09.202900 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert podName:49030eaa-0443-4db5-bfd2-3c4e17a3780e nodeName:}" failed. No retries permitted until 2025-12-11 08:54:11.202883862 +0000 UTC m=+1039.268299470 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fr7wqh" (UID: "49030eaa-0443-4db5-bfd2-3c4e17a3780e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.307672 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.307744 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2f78\" (UniqueName: \"kubernetes.io/projected/914206a5-adb8-4482-90f6-e55a294259c3-kube-api-access-c2f78\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.307770 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:09.308037 4629 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:09.308114 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs podName:914206a5-adb8-4482-90f6-e55a294259c3 nodeName:}" failed. No retries permitted until 2025-12-11 08:54:09.808069745 +0000 UTC m=+1037.873485363 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs") pod "openstack-operator-controller-manager-69b5bb5b-x6z9b" (UID: "914206a5-adb8-4482-90f6-e55a294259c3") : secret "webhook-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:09.308168 4629 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:09.308220 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs podName:914206a5-adb8-4482-90f6-e55a294259c3 nodeName:}" failed. No retries permitted until 2025-12-11 08:54:09.808209439 +0000 UTC m=+1037.873625047 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs") pod "openstack-operator-controller-manager-69b5bb5b-x6z9b" (UID: "914206a5-adb8-4482-90f6-e55a294259c3") : secret "metrics-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.450482 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2f78\" (UniqueName: \"kubernetes.io/projected/914206a5-adb8-4482-90f6-e55a294259c3-kube-api-access-c2f78\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.485647 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd"] Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.679918 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf"] Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.690833 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.701314 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-rzpkj" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.807570 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf"] Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.865524 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.865564 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.865618 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dvzt\" (UniqueName: \"kubernetes.io/projected/5af0ef89-5415-482d-bbe9-d97cb6da58d6-kube-api-access-5dvzt\") pod \"rabbitmq-cluster-operator-manager-668c99d594-kvzwf\" (UID: \"5af0ef89-5415-482d-bbe9-d97cb6da58d6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf" Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:09.865807 4629 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:09.865880 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs podName:914206a5-adb8-4482-90f6-e55a294259c3 nodeName:}" failed. No retries permitted until 2025-12-11 08:54:10.865863949 +0000 UTC m=+1038.931279547 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs") pod "openstack-operator-controller-manager-69b5bb5b-x6z9b" (UID: "914206a5-adb8-4482-90f6-e55a294259c3") : secret "metrics-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:09.866191 4629 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:09.866221 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs podName:914206a5-adb8-4482-90f6-e55a294259c3 nodeName:}" failed. No retries permitted until 2025-12-11 08:54:10.86621316 +0000 UTC m=+1038.931628768 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs") pod "openstack-operator-controller-manager-69b5bb5b-x6z9b" (UID: "914206a5-adb8-4482-90f6-e55a294259c3") : secret "webhook-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:09.967574 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dvzt\" (UniqueName: \"kubernetes.io/projected/5af0ef89-5415-482d-bbe9-d97cb6da58d6-kube-api-access-5dvzt\") pod \"rabbitmq-cluster-operator-manager-668c99d594-kvzwf\" (UID: \"5af0ef89-5415-482d-bbe9-d97cb6da58d6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:10.001120 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dvzt\" (UniqueName: \"kubernetes.io/projected/5af0ef89-5415-482d-bbe9-d97cb6da58d6-kube-api-access-5dvzt\") pod \"rabbitmq-cluster-operator-manager-668c99d594-kvzwf\" (UID: \"5af0ef89-5415-482d-bbe9-d97cb6da58d6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:10.263597 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-f8m45" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:10.276906 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:10.287740 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:10.751390 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx"] Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:10.910027 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert\") pod \"infra-operator-controller-manager-78d48bff9d-l59gh\" (UID: \"219ddd48-017d-4da4-b576-0b8f24bc40be\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:10.910475 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:10 crc kubenswrapper[4629]: I1211 08:54:10.910505 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:10.910300 4629 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:10.910608 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert podName:219ddd48-017d-4da4-b576-0b8f24bc40be nodeName:}" failed. No retries permitted until 2025-12-11 08:54:14.910564662 +0000 UTC m=+1042.975980340 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert") pod "infra-operator-controller-manager-78d48bff9d-l59gh" (UID: "219ddd48-017d-4da4-b576-0b8f24bc40be") : secret "infra-operator-webhook-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:10.910711 4629 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:10.910762 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs podName:914206a5-adb8-4482-90f6-e55a294259c3 nodeName:}" failed. No retries permitted until 2025-12-11 08:54:12.910747237 +0000 UTC m=+1040.976162835 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs") pod "openstack-operator-controller-manager-69b5bb5b-x6z9b" (UID: "914206a5-adb8-4482-90f6-e55a294259c3") : secret "webhook-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:10.910801 4629 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 08:54:10 crc kubenswrapper[4629]: E1211 08:54:10.910830 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs podName:914206a5-adb8-4482-90f6-e55a294259c3 nodeName:}" failed. No retries permitted until 2025-12-11 08:54:12.91081627 +0000 UTC m=+1040.976231908 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs") pod "openstack-operator-controller-manager-69b5bb5b-x6z9b" (UID: "914206a5-adb8-4482-90f6-e55a294259c3") : secret "metrics-server-cert" not found Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.242991 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd" event={"ID":"813d511c-6f11-40c8-a9c3-281e4909c88d","Type":"ContainerStarted","Data":"2c05c7569aedaa154602740ef4798a40671425add6c5a2730cd93f95f1170b02"} Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.244762 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx" event={"ID":"c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1","Type":"ContainerStarted","Data":"8d7d46d5e220de391a3188f026d5e5649e17083d4568fc5eb113ef2ef666a146"} Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.280239 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fr7wqh\" (UID: \"49030eaa-0443-4db5-bfd2-3c4e17a3780e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.280488 4629 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.280559 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert podName:49030eaa-0443-4db5-bfd2-3c4e17a3780e nodeName:}" failed. No retries permitted until 2025-12-11 08:54:15.280531888 +0000 UTC m=+1043.345947496 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fr7wqh" (UID: "49030eaa-0443-4db5-bfd2-3c4e17a3780e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.411241 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9"] Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.502198 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5"] Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.523530 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-f8m45"] Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.595662 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh"] Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.642655 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj"] Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.651004 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j"] Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.660721 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk"] Dec 11 08:54:11 crc kubenswrapper[4629]: W1211 08:54:11.671813 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc808480f_8564_4628_b9b9_7ee9700cbb91.slice/crio-032aedc68875750cdcfae9611fa5a1c121fd4e3082913c8768421dc7454ac77a WatchSource:0}: Error finding container 032aedc68875750cdcfae9611fa5a1c121fd4e3082913c8768421dc7454ac77a: Status 404 returned error can't find the container with id 032aedc68875750cdcfae9611fa5a1c121fd4e3082913c8768421dc7454ac77a Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.671885 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx"] Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.683787 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk"] Dec 11 08:54:11 crc kubenswrapper[4629]: W1211 08:54:11.692575 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a37b92a_08fc_47a9_8d55_97182286daea.slice/crio-1b33546de3bdcd267ee9093d195309f825ee154e9472c63ca539a38876bcd1a0 WatchSource:0}: Error finding container 1b33546de3bdcd267ee9093d195309f825ee154e9472c63ca539a38876bcd1a0: Status 404 returned error can't find the container with id 1b33546de3bdcd267ee9093d195309f825ee154e9472c63ca539a38876bcd1a0 Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.696456 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6"] Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.707091 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh"] Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.710011 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk"] Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.715800 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck"] Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.721905 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt"] Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.732653 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f9j6r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-tg2w6_openstack-operators(3daa4e0a-c804-4d70-a79c-5e4d5dd3a602): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.733235 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh"] Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.734140 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7fk84,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-mjhzk_openstack-operators(8c62910d-1b99-40df-be12-c3eb86c645aa): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.739816 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7fk84,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-mjhzk_openstack-operators(8c62910d-1b99-40df-be12-c3eb86c645aa): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.739978 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f9j6r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-tg2w6_openstack-operators(3daa4e0a-c804-4d70-a79c-5e4d5dd3a602): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.742154 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw"] Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.742252 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" podUID="3daa4e0a-c804-4d70-a79c-5e4d5dd3a602" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.742307 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" podUID="8c62910d-1b99-40df-be12-c3eb86c645aa" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.743807 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2mxjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-rmlvk_openstack-operators(bf9c412a-98bd-4152-b65a-637822c14b67): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.750474 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2mxjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-rmlvk_openstack-operators(bf9c412a-98bd-4152-b65a-637822c14b67): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.750568 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf"] Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.751644 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" podUID="bf9c412a-98bd-4152-b65a-637822c14b67" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.781315 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5dvzt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-kvzwf_openstack-operators(5af0ef89-5415-482d-bbe9-d97cb6da58d6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.785843 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf" podUID="5af0ef89-5415-482d-bbe9-d97cb6da58d6" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.791380 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v6fp2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-cptvh_openstack-operators(1150e5d5-ae4c-4800-a10d-9ba271074338): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.793366 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9n8nm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-8qz72_openstack-operators(a6ef2a86-805b-44ed-95bd-1c53e457b9d9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.794472 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mbcvx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-kqcvw_openstack-operators(94ec218d-3b08-463d-8225-f12cbcf0eb8d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.795940 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v6fp2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-cptvh_openstack-operators(1150e5d5-ae4c-4800-a10d-9ba271074338): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.796054 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9n8nm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-8qz72_openstack-operators(a6ef2a86-805b-44ed-95bd-1c53e457b9d9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.796807 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mbcvx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-kqcvw_openstack-operators(94ec218d-3b08-463d-8225-f12cbcf0eb8d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.797197 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" podUID="a6ef2a86-805b-44ed-95bd-1c53e457b9d9" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.797271 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" podUID="1150e5d5-ae4c-4800-a10d-9ba271074338" Dec 11 08:54:11 crc kubenswrapper[4629]: E1211 08:54:11.798401 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" podUID="94ec218d-3b08-463d-8225-f12cbcf0eb8d" Dec 11 08:54:11 crc kubenswrapper[4629]: I1211 08:54:11.812770 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72"] Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.284417 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj" event={"ID":"7833d1b6-8dbd-4afa-9042-755b5a9c383b","Type":"ContainerStarted","Data":"089a5de8c21d1a6ea0f31860f13696334ec9f74f51ebeaee661d6aa7ea8c9a47"} Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.294970 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk" event={"ID":"8e2cace8-e27c-4d40-b20e-dd16cb10a425","Type":"ContainerStarted","Data":"9e96c8e12ce8411ca143db547a0ad4e30b73399bb3fbd5ed13cdd6d6e434dd10"} Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.296883 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh" event={"ID":"99ac9552-d81c-47b9-a77c-c18fae357872","Type":"ContainerStarted","Data":"d6c5bbdb9524a7d7bca594c204df2ede3a1fd1bb36773a9ea2d9e29d7b458adb"} Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.297948 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-f8m45" event={"ID":"2382ec97-3dac-4285-bb95-fa1864b1bd8d","Type":"ContainerStarted","Data":"8637d7925c915b1a64b7c31e4b150c228f215282ea1dbc0bb4ae1969ae160a4c"} Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.299302 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt" event={"ID":"8a37b92a-08fc-47a9-8d55-97182286daea","Type":"ContainerStarted","Data":"1b33546de3bdcd267ee9093d195309f825ee154e9472c63ca539a38876bcd1a0"} Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.302777 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" event={"ID":"a6ef2a86-805b-44ed-95bd-1c53e457b9d9","Type":"ContainerStarted","Data":"018f24beda4bfa86c8a535701012ee9fc889964c20a5926db81a32179a131ac6"} Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.305392 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh" event={"ID":"0f40457d-36a2-4183-8434-9cb4688489bb","Type":"ContainerStarted","Data":"b7f4ad0616c9d64415ff48e2459ea43cdf4115d69ca9011f947e97d1563e2045"} Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.331332 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j" event={"ID":"64242f6c-6d25-4868-acd8-de067ecdeb7a","Type":"ContainerStarted","Data":"aeaac919fc87144750e38a4b3729344e4ad4110699e5653cbcf7a37317b41c0c"} Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.340247 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" event={"ID":"94ec218d-3b08-463d-8225-f12cbcf0eb8d","Type":"ContainerStarted","Data":"360b1960711232ae0ffbf98dae747931a934fc79f06d1687a85401bd3e3913fe"} Dec 11 08:54:12 crc kubenswrapper[4629]: E1211 08:54:12.344037 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" podUID="94ec218d-3b08-463d-8225-f12cbcf0eb8d" Dec 11 08:54:12 crc kubenswrapper[4629]: E1211 08:54:12.344061 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" podUID="a6ef2a86-805b-44ed-95bd-1c53e457b9d9" Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.345452 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" event={"ID":"3daa4e0a-c804-4d70-a79c-5e4d5dd3a602","Type":"ContainerStarted","Data":"022601e7fa06fe738e786f947e0610ffb4f55b440797fdaf40c9dbb13f2362c2"} Dec 11 08:54:12 crc kubenswrapper[4629]: E1211 08:54:12.348273 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" podUID="3daa4e0a-c804-4d70-a79c-5e4d5dd3a602" Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.352259 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" event={"ID":"1150e5d5-ae4c-4800-a10d-9ba271074338","Type":"ContainerStarted","Data":"27d00cda9b7fc942b6dc0cb704bf5e62a513419b2cfd620c72b20530fb2ee050"} Dec 11 08:54:12 crc kubenswrapper[4629]: E1211 08:54:12.355098 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" podUID="1150e5d5-ae4c-4800-a10d-9ba271074338" Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.369888 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" event={"ID":"8c62910d-1b99-40df-be12-c3eb86c645aa","Type":"ContainerStarted","Data":"95cdf086633b7638ea762849ce04492dad7a85205aacd3997d3726255dabc460"} Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.372957 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx" event={"ID":"c808480f-8564-4628-b9b9-7ee9700cbb91","Type":"ContainerStarted","Data":"032aedc68875750cdcfae9611fa5a1c121fd4e3082913c8768421dc7454ac77a"} Dec 11 08:54:12 crc kubenswrapper[4629]: E1211 08:54:12.374135 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" podUID="8c62910d-1b99-40df-be12-c3eb86c645aa" Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.379730 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck" event={"ID":"8ebecdcb-ee18-4285-8b09-65c8859dc77e","Type":"ContainerStarted","Data":"aef8b8eb6aad2d21ad8ebf493dc5fa362fdf6e8d5879cce4d093be88264db5d4"} Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.388325 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf" event={"ID":"5af0ef89-5415-482d-bbe9-d97cb6da58d6","Type":"ContainerStarted","Data":"58d4783d0532c9d08818482c103beea8473f1d1b5fc1c2a1a292530e292bf496"} Dec 11 08:54:12 crc kubenswrapper[4629]: E1211 08:54:12.389646 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf" podUID="5af0ef89-5415-482d-bbe9-d97cb6da58d6" Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.392277 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" event={"ID":"bf9c412a-98bd-4152-b65a-637822c14b67","Type":"ContainerStarted","Data":"49f85c5c339ef1e1723b86eccf240f2ac1fe17f47072980a7ded3cc577bb0894"} Dec 11 08:54:12 crc kubenswrapper[4629]: E1211 08:54:12.398656 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" podUID="bf9c412a-98bd-4152-b65a-637822c14b67" Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.398820 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5" event={"ID":"bf19951f-862f-4b74-92ab-63cdede5be6d","Type":"ContainerStarted","Data":"d18ae51d6eb3af3c61c278e73fd87c5d6c55776f6c93b6c73d40ab9cc63eb96b"} Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.408039 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9" event={"ID":"39798d73-c400-4797-b5ac-e402d17a1df1","Type":"ContainerStarted","Data":"972bf0f043930d94247c690aabd9f6a32b4148566a9b0b1f674770396effa890"} Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.916657 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:12 crc kubenswrapper[4629]: I1211 08:54:12.916721 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:12 crc kubenswrapper[4629]: E1211 08:54:12.916911 4629 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 08:54:12 crc kubenswrapper[4629]: E1211 08:54:12.916969 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs podName:914206a5-adb8-4482-90f6-e55a294259c3 nodeName:}" failed. No retries permitted until 2025-12-11 08:54:16.91695128 +0000 UTC m=+1044.982366888 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs") pod "openstack-operator-controller-manager-69b5bb5b-x6z9b" (UID: "914206a5-adb8-4482-90f6-e55a294259c3") : secret "webhook-server-cert" not found Dec 11 08:54:12 crc kubenswrapper[4629]: E1211 08:54:12.917391 4629 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 08:54:12 crc kubenswrapper[4629]: E1211 08:54:12.917425 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs podName:914206a5-adb8-4482-90f6-e55a294259c3 nodeName:}" failed. No retries permitted until 2025-12-11 08:54:16.917414484 +0000 UTC m=+1044.982830092 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs") pod "openstack-operator-controller-manager-69b5bb5b-x6z9b" (UID: "914206a5-adb8-4482-90f6-e55a294259c3") : secret "metrics-server-cert" not found Dec 11 08:54:13 crc kubenswrapper[4629]: E1211 08:54:13.443295 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" podUID="94ec218d-3b08-463d-8225-f12cbcf0eb8d" Dec 11 08:54:13 crc kubenswrapper[4629]: E1211 08:54:13.443587 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" podUID="1150e5d5-ae4c-4800-a10d-9ba271074338" Dec 11 08:54:13 crc kubenswrapper[4629]: E1211 08:54:13.445428 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" podUID="3daa4e0a-c804-4d70-a79c-5e4d5dd3a602" Dec 11 08:54:13 crc kubenswrapper[4629]: E1211 08:54:13.447090 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" podUID="a6ef2a86-805b-44ed-95bd-1c53e457b9d9" Dec 11 08:54:13 crc kubenswrapper[4629]: E1211 08:54:13.448644 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf" podUID="5af0ef89-5415-482d-bbe9-d97cb6da58d6" Dec 11 08:54:13 crc kubenswrapper[4629]: E1211 08:54:13.451093 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" podUID="8c62910d-1b99-40df-be12-c3eb86c645aa" Dec 11 08:54:13 crc kubenswrapper[4629]: E1211 08:54:13.453911 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" podUID="bf9c412a-98bd-4152-b65a-637822c14b67" Dec 11 08:54:14 crc kubenswrapper[4629]: I1211 08:54:14.984062 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert\") pod \"infra-operator-controller-manager-78d48bff9d-l59gh\" (UID: \"219ddd48-017d-4da4-b576-0b8f24bc40be\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:14 crc kubenswrapper[4629]: E1211 08:54:14.984315 4629 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 08:54:14 crc kubenswrapper[4629]: E1211 08:54:14.984375 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert podName:219ddd48-017d-4da4-b576-0b8f24bc40be nodeName:}" failed. No retries permitted until 2025-12-11 08:54:22.984356884 +0000 UTC m=+1051.049772492 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert") pod "infra-operator-controller-manager-78d48bff9d-l59gh" (UID: "219ddd48-017d-4da4-b576-0b8f24bc40be") : secret "infra-operator-webhook-server-cert" not found Dec 11 08:54:15 crc kubenswrapper[4629]: I1211 08:54:15.288355 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fr7wqh\" (UID: \"49030eaa-0443-4db5-bfd2-3c4e17a3780e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:15 crc kubenswrapper[4629]: E1211 08:54:15.289327 4629 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:54:15 crc kubenswrapper[4629]: E1211 08:54:15.289403 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert podName:49030eaa-0443-4db5-bfd2-3c4e17a3780e nodeName:}" failed. No retries permitted until 2025-12-11 08:54:23.289383422 +0000 UTC m=+1051.354799030 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fr7wqh" (UID: "49030eaa-0443-4db5-bfd2-3c4e17a3780e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:54:17 crc kubenswrapper[4629]: I1211 08:54:17.016643 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:17 crc kubenswrapper[4629]: I1211 08:54:17.016717 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:17 crc kubenswrapper[4629]: E1211 08:54:17.016861 4629 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 08:54:17 crc kubenswrapper[4629]: E1211 08:54:17.016950 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs podName:914206a5-adb8-4482-90f6-e55a294259c3 nodeName:}" failed. No retries permitted until 2025-12-11 08:54:25.016927284 +0000 UTC m=+1053.082342942 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs") pod "openstack-operator-controller-manager-69b5bb5b-x6z9b" (UID: "914206a5-adb8-4482-90f6-e55a294259c3") : secret "metrics-server-cert" not found Dec 11 08:54:17 crc kubenswrapper[4629]: E1211 08:54:17.016953 4629 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 08:54:17 crc kubenswrapper[4629]: E1211 08:54:17.017022 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs podName:914206a5-adb8-4482-90f6-e55a294259c3 nodeName:}" failed. No retries permitted until 2025-12-11 08:54:25.017001087 +0000 UTC m=+1053.082416775 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs") pod "openstack-operator-controller-manager-69b5bb5b-x6z9b" (UID: "914206a5-adb8-4482-90f6-e55a294259c3") : secret "webhook-server-cert" not found Dec 11 08:54:23 crc kubenswrapper[4629]: I1211 08:54:23.033951 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert\") pod \"infra-operator-controller-manager-78d48bff9d-l59gh\" (UID: \"219ddd48-017d-4da4-b576-0b8f24bc40be\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:23 crc kubenswrapper[4629]: I1211 08:54:23.044625 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/219ddd48-017d-4da4-b576-0b8f24bc40be-cert\") pod \"infra-operator-controller-manager-78d48bff9d-l59gh\" (UID: \"219ddd48-017d-4da4-b576-0b8f24bc40be\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:23 crc kubenswrapper[4629]: I1211 08:54:23.051196 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:23 crc kubenswrapper[4629]: I1211 08:54:23.339704 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fr7wqh\" (UID: \"49030eaa-0443-4db5-bfd2-3c4e17a3780e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:23 crc kubenswrapper[4629]: E1211 08:54:23.339912 4629 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:54:23 crc kubenswrapper[4629]: E1211 08:54:23.340012 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert podName:49030eaa-0443-4db5-bfd2-3c4e17a3780e nodeName:}" failed. No retries permitted until 2025-12-11 08:54:39.33998438 +0000 UTC m=+1067.405399988 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fr7wqh" (UID: "49030eaa-0443-4db5-bfd2-3c4e17a3780e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:54:25 crc kubenswrapper[4629]: I1211 08:54:25.064634 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:25 crc kubenswrapper[4629]: I1211 08:54:25.065027 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:25 crc kubenswrapper[4629]: I1211 08:54:25.077273 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-metrics-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:25 crc kubenswrapper[4629]: I1211 08:54:25.079365 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/914206a5-adb8-4482-90f6-e55a294259c3-webhook-certs\") pod \"openstack-operator-controller-manager-69b5bb5b-x6z9b\" (UID: \"914206a5-adb8-4482-90f6-e55a294259c3\") " pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:25 crc kubenswrapper[4629]: I1211 08:54:25.335398 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:25 crc kubenswrapper[4629]: E1211 08:54:25.701056 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87" Dec 11 08:54:25 crc kubenswrapper[4629]: E1211 08:54:25.701258 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8mz5m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-967d97867-qxsl5_openstack-operators(bf19951f-862f-4b74-92ab-63cdede5be6d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:54:26 crc kubenswrapper[4629]: E1211 08:54:26.289432 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3" Dec 11 08:54:26 crc kubenswrapper[4629]: E1211 08:54:26.289985 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zhv7n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-6c677c69b-p8lh9_openstack-operators(39798d73-c400-4797-b5ac-e402d17a1df1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:54:27 crc kubenswrapper[4629]: E1211 08:54:27.997015 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 11 08:54:27 crc kubenswrapper[4629]: E1211 08:54:27.997531 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f2llp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-f8m45_openstack-operators(2382ec97-3dac-4285-bb95-fa1864b1bd8d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:54:30 crc kubenswrapper[4629]: E1211 08:54:30.071082 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027" Dec 11 08:54:30 crc kubenswrapper[4629]: E1211 08:54:30.071309 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9nsxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-xdp6j_openstack-operators(64242f6c-6d25-4868-acd8-de067ecdeb7a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:54:33 crc kubenswrapper[4629]: E1211 08:54:33.564370 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 11 08:54:33 crc kubenswrapper[4629]: E1211 08:54:33.564576 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cvxh8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-8mtlt_openstack-operators(8a37b92a-08fc-47a9-8d55-97182286daea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:54:33 crc kubenswrapper[4629]: I1211 08:54:33.575857 4629 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 08:54:34 crc kubenswrapper[4629]: E1211 08:54:34.206462 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a" Dec 11 08:54:34 crc kubenswrapper[4629]: E1211 08:54:34.207178 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hl9nj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-697fb699cf-mhqnx_openstack-operators(c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:54:34 crc kubenswrapper[4629]: I1211 08:54:34.694954 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b"] Dec 11 08:54:35 crc kubenswrapper[4629]: E1211 08:54:35.702086 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 11 08:54:35 crc kubenswrapper[4629]: E1211 08:54:35.702277 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6zsj8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-p9wjh_openstack-operators(99ac9552-d81c-47b9-a77c-c18fae357872): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:54:39 crc kubenswrapper[4629]: I1211 08:54:39.371030 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fr7wqh\" (UID: \"49030eaa-0443-4db5-bfd2-3c4e17a3780e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:39 crc kubenswrapper[4629]: I1211 08:54:39.381702 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49030eaa-0443-4db5-bfd2-3c4e17a3780e-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fr7wqh\" (UID: \"49030eaa-0443-4db5-bfd2-3c4e17a3780e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:39 crc kubenswrapper[4629]: I1211 08:54:39.386982 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:42 crc kubenswrapper[4629]: I1211 08:54:42.345740 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh"] Dec 11 08:54:42 crc kubenswrapper[4629]: I1211 08:54:42.688960 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" event={"ID":"914206a5-adb8-4482-90f6-e55a294259c3","Type":"ContainerStarted","Data":"935710e8253acc958d49e2a4f301dd0391f0533617d5fb608e8e1fa9d40a5a19"} Dec 11 08:54:43 crc kubenswrapper[4629]: E1211 08:54:43.962685 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 11 08:54:43 crc kubenswrapper[4629]: E1211 08:54:43.963941 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mbcvx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-kqcvw_openstack-operators(94ec218d-3b08-463d-8225-f12cbcf0eb8d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:54:44 crc kubenswrapper[4629]: I1211 08:54:44.719978 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" event={"ID":"219ddd48-017d-4da4-b576-0b8f24bc40be","Type":"ContainerStarted","Data":"3224da6870aca1a1e744d4b4f1ffb80e12aaae9a216e33e710d44af4008d5678"} Dec 11 08:54:44 crc kubenswrapper[4629]: I1211 08:54:44.758584 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh"] Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.784768 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj" event={"ID":"7833d1b6-8dbd-4afa-9042-755b5a9c383b","Type":"ContainerStarted","Data":"b06674dda34c78877fdb8c3445793370d5a7ce544855d6e6eefec939c997d43e"} Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.807660 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" event={"ID":"1150e5d5-ae4c-4800-a10d-9ba271074338","Type":"ContainerStarted","Data":"7b9215efcbec1bd556d0843a522ac862d041b9959170e1e8b13b57ee77bd5c83"} Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.809484 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" event={"ID":"8c62910d-1b99-40df-be12-c3eb86c645aa","Type":"ContainerStarted","Data":"3b421c41b2ce49f9d48bac0b04bab239f9c4f441200d6931f8eb3a0ad69ffeaf"} Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.811686 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk" event={"ID":"8e2cace8-e27c-4d40-b20e-dd16cb10a425","Type":"ContainerStarted","Data":"eb5c9e0361a5bfcfea2825a9658008540410d1466d200a8d57a5e385cedff228"} Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.813150 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd" event={"ID":"813d511c-6f11-40c8-a9c3-281e4909c88d","Type":"ContainerStarted","Data":"05ce50bc9a56237c2a39a68fe842d0e95ed83d1f46667812a2fc325cb2f794a9"} Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.814407 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" event={"ID":"a6ef2a86-805b-44ed-95bd-1c53e457b9d9","Type":"ContainerStarted","Data":"924387f602b1cac427353dbf00717180e870442566c38b91de3164aeea99072a"} Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.815416 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" event={"ID":"914206a5-adb8-4482-90f6-e55a294259c3","Type":"ContainerStarted","Data":"58fa67148e11ed7459b29368ae53e9a9a4d4a19cf22aec2c1baeb0a106d757b8"} Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.816221 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.817664 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx" event={"ID":"c808480f-8564-4628-b9b9-7ee9700cbb91","Type":"ContainerStarted","Data":"a0b55735e97e4ea790afced70d806e671b5b978cabe50244b4650e61c36ab9f1"} Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.829320 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh" event={"ID":"0f40457d-36a2-4183-8434-9cb4688489bb","Type":"ContainerStarted","Data":"a42ef5080be8bfa9c61bce6a5908218bd0672d477d863e2454c9a426f27514a0"} Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.831632 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" event={"ID":"49030eaa-0443-4db5-bfd2-3c4e17a3780e","Type":"ContainerStarted","Data":"3bc8c5e43817f03dd9663b2836918d14ac50ecb85549927cad704fb72c0db606"} Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.842666 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck" event={"ID":"8ebecdcb-ee18-4285-8b09-65c8859dc77e","Type":"ContainerStarted","Data":"995ff9a3db109918649d44494f731f164c23271d1f3e0ddad57c8771e3e41969"} Dec 11 08:54:45 crc kubenswrapper[4629]: I1211 08:54:45.873330 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" podStartSLOduration=37.873298991 podStartE2EDuration="37.873298991s" podCreationTimestamp="2025-12-11 08:54:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:54:45.857314859 +0000 UTC m=+1073.922730467" watchObservedRunningTime="2025-12-11 08:54:45.873298991 +0000 UTC m=+1073.938714599" Dec 11 08:54:47 crc kubenswrapper[4629]: I1211 08:54:47.862380 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" event={"ID":"3daa4e0a-c804-4d70-a79c-5e4d5dd3a602","Type":"ContainerStarted","Data":"a3ea545bca047fb0661f8b68ce0b172417d701a323a23da12b73f31d2b826a06"} Dec 11 08:54:47 crc kubenswrapper[4629]: I1211 08:54:47.866023 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" event={"ID":"bf9c412a-98bd-4152-b65a-637822c14b67","Type":"ContainerStarted","Data":"f45bbd05a0cf56418205565c29c0ba1acf9c799b87137139cd547163288ce560"} Dec 11 08:54:47 crc kubenswrapper[4629]: I1211 08:54:47.870524 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf" event={"ID":"5af0ef89-5415-482d-bbe9-d97cb6da58d6","Type":"ContainerStarted","Data":"8a5f79b99e83dd897b146b03fc3529196a35622e80adb3e7ed58fcbad0791715"} Dec 11 08:54:47 crc kubenswrapper[4629]: I1211 08:54:47.894671 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvzwf" podStartSLOduration=6.123140131 podStartE2EDuration="38.894647629s" podCreationTimestamp="2025-12-11 08:54:09 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.781055984 +0000 UTC m=+1039.846471592" lastFinishedPulling="2025-12-11 08:54:44.552563482 +0000 UTC m=+1072.617979090" observedRunningTime="2025-12-11 08:54:47.890118247 +0000 UTC m=+1075.955533855" watchObservedRunningTime="2025-12-11 08:54:47.894647629 +0000 UTC m=+1075.960063237" Dec 11 08:54:53 crc kubenswrapper[4629]: E1211 08:54:53.037505 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-f8m45" podUID="2382ec97-3dac-4285-bb95-fa1864b1bd8d" Dec 11 08:54:53 crc kubenswrapper[4629]: E1211 08:54:53.148340 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx" podUID="c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1" Dec 11 08:54:53 crc kubenswrapper[4629]: E1211 08:54:53.225284 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt" podUID="8a37b92a-08fc-47a9-8d55-97182286daea" Dec 11 08:54:53 crc kubenswrapper[4629]: E1211 08:54:53.247458 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j" podUID="64242f6c-6d25-4868-acd8-de067ecdeb7a" Dec 11 08:54:53 crc kubenswrapper[4629]: E1211 08:54:53.494497 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5" podUID="bf19951f-862f-4b74-92ab-63cdede5be6d" Dec 11 08:54:53 crc kubenswrapper[4629]: E1211 08:54:53.599153 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" podUID="94ec218d-3b08-463d-8225-f12cbcf0eb8d" Dec 11 08:54:53 crc kubenswrapper[4629]: E1211 08:54:53.862695 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9" podUID="39798d73-c400-4797-b5ac-e402d17a1df1" Dec 11 08:54:53 crc kubenswrapper[4629]: E1211 08:54:53.939800 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh" podUID="99ac9552-d81c-47b9-a77c-c18fae357872" Dec 11 08:54:53 crc kubenswrapper[4629]: I1211 08:54:53.940977 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt" event={"ID":"8a37b92a-08fc-47a9-8d55-97182286daea","Type":"ContainerStarted","Data":"a7e1ee4e33476de9369dfcc0344aee07d9d8c35222e68b11786b5589ae2acad6"} Dec 11 08:54:53 crc kubenswrapper[4629]: I1211 08:54:53.955531 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j" event={"ID":"64242f6c-6d25-4868-acd8-de067ecdeb7a","Type":"ContainerStarted","Data":"a3cfc891a7b69ff956ff3c2194bbb76158a23593caf3985d3c7e200afcea6d3f"} Dec 11 08:54:53 crc kubenswrapper[4629]: I1211 08:54:53.981255 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx" event={"ID":"c808480f-8564-4628-b9b9-7ee9700cbb91","Type":"ContainerStarted","Data":"e435306a234172773c4ba7a2b5e0b5cd07d3af9721b0aadcb0f8ee705323b641"} Dec 11 08:54:53 crc kubenswrapper[4629]: I1211 08:54:53.982185 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:53.999214 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.004474 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-f8m45" event={"ID":"2382ec97-3dac-4285-bb95-fa1864b1bd8d","Type":"ContainerStarted","Data":"78bdf83d730fa8ea99f2c4fd1b07668c8445c9010e56422c9f55c2f2438c2525"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.014366 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9" event={"ID":"39798d73-c400-4797-b5ac-e402d17a1df1","Type":"ContainerStarted","Data":"1d2bdb2a752a1088909463b8a5fea3f2fc2914e4bac143c3fd9c7a97fbb2afd9"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.017261 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" event={"ID":"a6ef2a86-805b-44ed-95bd-1c53e457b9d9","Type":"ContainerStarted","Data":"fc240de88c177579cc9674e5f6a30fd4bae764bf1ef84e4f8c8caee9ff32de72"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.017839 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.021976 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.031478 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj" event={"ID":"7833d1b6-8dbd-4afa-9042-755b5a9c383b","Type":"ContainerStarted","Data":"7dc8d50394b81d401039bf548cfc39e01aae01243a859d12ef77824c414f5947"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.032418 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.033704 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5" event={"ID":"bf19951f-862f-4b74-92ab-63cdede5be6d","Type":"ContainerStarted","Data":"3e6c117c74d53baa649cd8511dfb8d45cd3a60c8049f8ee75a54605a9b104a4c"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.042093 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.059892 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh" event={"ID":"0f40457d-36a2-4183-8434-9cb4688489bb","Type":"ContainerStarted","Data":"3324d557b9197551f6c343876f8ca444ba6ca96f6590ee914156328f5f14c2f9"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.060202 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.061954 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" event={"ID":"49030eaa-0443-4db5-bfd2-3c4e17a3780e","Type":"ContainerStarted","Data":"281c669959fa7579d97fd71afbd7ac95bd2095808614be9fe577980b139ee695"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.061993 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" event={"ID":"49030eaa-0443-4db5-bfd2-3c4e17a3780e","Type":"ContainerStarted","Data":"368f5760a8b622a67d60c17e924c4525fb0b7ad13d30ffeb57b226297f40872b"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.062736 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.064973 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-f5ftx" podStartSLOduration=6.051009709 podStartE2EDuration="47.064951269s" podCreationTimestamp="2025-12-11 08:54:07 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.695949461 +0000 UTC m=+1039.761365069" lastFinishedPulling="2025-12-11 08:54:52.709891021 +0000 UTC m=+1080.775306629" observedRunningTime="2025-12-11 08:54:54.05989191 +0000 UTC m=+1082.125307518" watchObservedRunningTime="2025-12-11 08:54:54.064951269 +0000 UTC m=+1082.130366877" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.080035 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" event={"ID":"94ec218d-3b08-463d-8225-f12cbcf0eb8d","Type":"ContainerStarted","Data":"8f049dd2c52028e386a4f19d7e0d14df63ff01b93f036dde8716a9d77d156ab7"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.085076 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh" Dec 11 08:54:54 crc kubenswrapper[4629]: E1211 08:54:54.088240 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" podUID="94ec218d-3b08-463d-8225-f12cbcf0eb8d" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.114091 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" event={"ID":"8c62910d-1b99-40df-be12-c3eb86c645aa","Type":"ContainerStarted","Data":"b409c520d8808ca4b75ad27e1ba1a36b07b3257f5174455fbbfddb039e3dd050"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.114952 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.126099 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.132315 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" event={"ID":"219ddd48-017d-4da4-b576-0b8f24bc40be","Type":"ContainerStarted","Data":"e01f984255e77862285658fe6644e76f047bac3d5181aa98a6ed82f386785740"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.132356 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" event={"ID":"219ddd48-017d-4da4-b576-0b8f24bc40be","Type":"ContainerStarted","Data":"59d845336315c0fad2aa92df7155cbdc4cef672a50042d5c0a7a74cf39ba1c90"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.133052 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.139973 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-8qz72" podStartSLOduration=6.35870538 podStartE2EDuration="47.139957394s" podCreationTimestamp="2025-12-11 08:54:07 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.792180393 +0000 UTC m=+1039.857596001" lastFinishedPulling="2025-12-11 08:54:52.573432407 +0000 UTC m=+1080.638848015" observedRunningTime="2025-12-11 08:54:54.136912598 +0000 UTC m=+1082.202328206" watchObservedRunningTime="2025-12-11 08:54:54.139957394 +0000 UTC m=+1082.205373002" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.170299 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd" event={"ID":"813d511c-6f11-40c8-a9c3-281e4909c88d","Type":"ContainerStarted","Data":"65f29f38496fd90646069d11c77889c26997952210af98b5b4f298c4ed896d59"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.171393 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.183312 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.240399 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx" event={"ID":"c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1","Type":"ContainerStarted","Data":"568929080b07ba501dbcb4203a06ecf132554dba61db9f4981c1f72756a80a3d"} Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.254584 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" podStartSLOduration=39.680110076 podStartE2EDuration="48.254561072s" podCreationTimestamp="2025-12-11 08:54:06 +0000 UTC" firstStartedPulling="2025-12-11 08:54:43.962749263 +0000 UTC m=+1072.028164911" lastFinishedPulling="2025-12-11 08:54:52.537200299 +0000 UTC m=+1080.602615907" observedRunningTime="2025-12-11 08:54:54.236464394 +0000 UTC m=+1082.301880002" watchObservedRunningTime="2025-12-11 08:54:54.254561072 +0000 UTC m=+1082.319976680" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.329171 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kq7vj" podStartSLOduration=7.334243732 podStartE2EDuration="48.329153345s" podCreationTimestamp="2025-12-11 08:54:06 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.628030058 +0000 UTC m=+1039.693445666" lastFinishedPulling="2025-12-11 08:54:52.622939671 +0000 UTC m=+1080.688355279" observedRunningTime="2025-12-11 08:54:54.324078496 +0000 UTC m=+1082.389494104" watchObservedRunningTime="2025-12-11 08:54:54.329153345 +0000 UTC m=+1082.394568953" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.563513 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" podStartSLOduration=39.846509929 podStartE2EDuration="47.563498443s" podCreationTimestamp="2025-12-11 08:54:07 +0000 UTC" firstStartedPulling="2025-12-11 08:54:44.820212925 +0000 UTC m=+1072.885628523" lastFinishedPulling="2025-12-11 08:54:52.537201429 +0000 UTC m=+1080.602617037" observedRunningTime="2025-12-11 08:54:54.561570292 +0000 UTC m=+1082.626985900" watchObservedRunningTime="2025-12-11 08:54:54.563498443 +0000 UTC m=+1082.628914051" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.564916 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g5zpd" podStartSLOduration=6.121761071 podStartE2EDuration="48.564836355s" podCreationTimestamp="2025-12-11 08:54:06 +0000 UTC" firstStartedPulling="2025-12-11 08:54:10.234828534 +0000 UTC m=+1038.300244142" lastFinishedPulling="2025-12-11 08:54:52.677903818 +0000 UTC m=+1080.743319426" observedRunningTime="2025-12-11 08:54:54.373209538 +0000 UTC m=+1082.438625146" watchObservedRunningTime="2025-12-11 08:54:54.564836355 +0000 UTC m=+1082.630251953" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.731704 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9m9lh" podStartSLOduration=7.677136869 podStartE2EDuration="48.731686154s" podCreationTimestamp="2025-12-11 08:54:06 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.639687705 +0000 UTC m=+1039.705103313" lastFinishedPulling="2025-12-11 08:54:52.69423699 +0000 UTC m=+1080.759652598" observedRunningTime="2025-12-11 08:54:54.715910339 +0000 UTC m=+1082.781325967" watchObservedRunningTime="2025-12-11 08:54:54.731686154 +0000 UTC m=+1082.797101762" Dec 11 08:54:54 crc kubenswrapper[4629]: I1211 08:54:54.779392 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjhzk" podStartSLOduration=7.821213383 podStartE2EDuration="48.779370911s" podCreationTimestamp="2025-12-11 08:54:06 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.734054628 +0000 UTC m=+1039.799470236" lastFinishedPulling="2025-12-11 08:54:52.692212146 +0000 UTC m=+1080.757627764" observedRunningTime="2025-12-11 08:54:54.76308847 +0000 UTC m=+1082.828504078" watchObservedRunningTime="2025-12-11 08:54:54.779370911 +0000 UTC m=+1082.844786519" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.221877 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh" event={"ID":"99ac9552-d81c-47b9-a77c-c18fae357872","Type":"ContainerStarted","Data":"8de774eed2e312484564f2df0893ecde235e943e5206c730d28bb07b0e468219"} Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.224133 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt" event={"ID":"8a37b92a-08fc-47a9-8d55-97182286daea","Type":"ContainerStarted","Data":"3771c3598810c329f43ea24de26c034f7f1d212f5a90312da21e3efcac4ef358"} Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.224234 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.227311 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck" event={"ID":"8ebecdcb-ee18-4285-8b09-65c8859dc77e","Type":"ContainerStarted","Data":"869a42c238520ca4b0068a59fe4ced5dcefb915fc8e0e34853825366d5a5fd33"} Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.228071 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.230701 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j" event={"ID":"64242f6c-6d25-4868-acd8-de067ecdeb7a","Type":"ContainerStarted","Data":"5196fdb3f10e212d44b3979c391e8737d7ecb68fee77a911799c2cb17d029bda"} Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.230898 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.232429 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" event={"ID":"3daa4e0a-c804-4d70-a79c-5e4d5dd3a602","Type":"ContainerStarted","Data":"6030126dbaa01a4590f781d73541bc475d57582a3f52db57a59d05df27c2a7a0"} Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.233270 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.234367 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.235523 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" event={"ID":"bf9c412a-98bd-4152-b65a-637822c14b67","Type":"ContainerStarted","Data":"768e770c0ece634a148734499f49b97814f5bedd072db545bd8ffa0bfdae2456"} Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.236199 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.238790 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" event={"ID":"1150e5d5-ae4c-4800-a10d-9ba271074338","Type":"ContainerStarted","Data":"aca33c342f5780952e24bb17439ad43b488af0768a43f289b1d3ca70df735497"} Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.239089 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.241046 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.241094 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.242510 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk" event={"ID":"8e2cace8-e27c-4d40-b20e-dd16cb10a425","Type":"ContainerStarted","Data":"7652eef199bdee456ab753e7ba44f9114aaaeb3136dfb39f0db39db41409b1ba"} Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.242532 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.246268 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.249377 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.278630 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-tg2w6" podStartSLOduration=7.26285706 podStartE2EDuration="48.278613287s" podCreationTimestamp="2025-12-11 08:54:07 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.732501599 +0000 UTC m=+1039.797917207" lastFinishedPulling="2025-12-11 08:54:52.748257826 +0000 UTC m=+1080.813673434" observedRunningTime="2025-12-11 08:54:55.273555687 +0000 UTC m=+1083.338971295" watchObservedRunningTime="2025-12-11 08:54:55.278613287 +0000 UTC m=+1083.344028895" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.311988 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-rmlvk" podStartSLOduration=8.341924332 podStartE2EDuration="49.311963784s" podCreationTimestamp="2025-12-11 08:54:06 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.74369344 +0000 UTC m=+1039.809109058" lastFinishedPulling="2025-12-11 08:54:52.713732902 +0000 UTC m=+1080.779148510" observedRunningTime="2025-12-11 08:54:55.302796516 +0000 UTC m=+1083.368212124" watchObservedRunningTime="2025-12-11 08:54:55.311963784 +0000 UTC m=+1083.377379392" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.334408 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-cptvh" podStartSLOduration=7.427951824 podStartE2EDuration="48.334390968s" podCreationTimestamp="2025-12-11 08:54:07 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.791201522 +0000 UTC m=+1039.856617120" lastFinishedPulling="2025-12-11 08:54:52.697640646 +0000 UTC m=+1080.763056264" observedRunningTime="2025-12-11 08:54:55.33351402 +0000 UTC m=+1083.398929628" watchObservedRunningTime="2025-12-11 08:54:55.334390968 +0000 UTC m=+1083.399806576" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.365202 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-69b5bb5b-x6z9b" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.392408 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j" podStartSLOduration=6.478852954 podStartE2EDuration="49.392391389s" podCreationTimestamp="2025-12-11 08:54:06 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.628475083 +0000 UTC m=+1039.693890681" lastFinishedPulling="2025-12-11 08:54:54.542013508 +0000 UTC m=+1082.607429116" observedRunningTime="2025-12-11 08:54:55.38764065 +0000 UTC m=+1083.453056258" watchObservedRunningTime="2025-12-11 08:54:55.392391389 +0000 UTC m=+1083.457806997" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.463894 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8q2kk" podStartSLOduration=7.459422851 podStartE2EDuration="48.463868773s" podCreationTimestamp="2025-12-11 08:54:07 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.62710895 +0000 UTC m=+1039.692524558" lastFinishedPulling="2025-12-11 08:54:52.631554862 +0000 UTC m=+1080.696970480" observedRunningTime="2025-12-11 08:54:55.422586287 +0000 UTC m=+1083.488001895" watchObservedRunningTime="2025-12-11 08:54:55.463868773 +0000 UTC m=+1083.529284391" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.464484 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt" podStartSLOduration=6.57653078 podStartE2EDuration="49.464478352s" podCreationTimestamp="2025-12-11 08:54:06 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.730155926 +0000 UTC m=+1039.795571534" lastFinishedPulling="2025-12-11 08:54:54.618103498 +0000 UTC m=+1082.683519106" observedRunningTime="2025-12-11 08:54:55.450744791 +0000 UTC m=+1083.516160409" watchObservedRunningTime="2025-12-11 08:54:55.464478352 +0000 UTC m=+1083.529893960" Dec 11 08:54:55 crc kubenswrapper[4629]: I1211 08:54:55.478809 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-gtqck" podStartSLOduration=7.479873594 podStartE2EDuration="48.478787992s" podCreationTimestamp="2025-12-11 08:54:07 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.730466135 +0000 UTC m=+1039.795881743" lastFinishedPulling="2025-12-11 08:54:52.729380533 +0000 UTC m=+1080.794796141" observedRunningTime="2025-12-11 08:54:55.478462371 +0000 UTC m=+1083.543877999" watchObservedRunningTime="2025-12-11 08:54:55.478787992 +0000 UTC m=+1083.544203600" Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.274183 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-f8m45" event={"ID":"2382ec97-3dac-4285-bb95-fa1864b1bd8d","Type":"ContainerStarted","Data":"9208793f65b857d54bcb6ec484a905834447fc376384a3d70e8d5ee7a99eabc3"} Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.275418 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-f8m45" Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.281533 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9" event={"ID":"39798d73-c400-4797-b5ac-e402d17a1df1","Type":"ContainerStarted","Data":"bec02977067e8b7036c26b302b59387c5f21af5e54c74e4882b803e8ecb2aaa2"} Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.282451 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9" Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.285226 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx" event={"ID":"c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1","Type":"ContainerStarted","Data":"cb152da08e422c7e311299b0f14e071566da5798cd639c95aa4c0cb16456b903"} Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.285935 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx" Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.295134 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5" event={"ID":"bf19951f-862f-4b74-92ab-63cdede5be6d","Type":"ContainerStarted","Data":"8eb3cd4ce1c3fc862db5ac23bc36b7fb1298bb3367eede212978d7da777dfbac"} Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.295732 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5" Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.307012 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh" event={"ID":"99ac9552-d81c-47b9-a77c-c18fae357872","Type":"ContainerStarted","Data":"60e11d8a4e09bed690ba7fd8a74664817c71a6f0e553f2c5ea6063b83c3f4973"} Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.307055 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh" Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.308923 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-f8m45" podStartSLOduration=5.450186354 podStartE2EDuration="49.308911237s" podCreationTimestamp="2025-12-11 08:54:07 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.624412406 +0000 UTC m=+1039.689828004" lastFinishedPulling="2025-12-11 08:54:55.483137279 +0000 UTC m=+1083.548552887" observedRunningTime="2025-12-11 08:54:56.305062886 +0000 UTC m=+1084.370478494" watchObservedRunningTime="2025-12-11 08:54:56.308911237 +0000 UTC m=+1084.374326845" Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.345724 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5" podStartSLOduration=6.936201484 podStartE2EDuration="50.345700031s" podCreationTimestamp="2025-12-11 08:54:06 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.606193193 +0000 UTC m=+1039.671608811" lastFinishedPulling="2025-12-11 08:54:55.01569175 +0000 UTC m=+1083.081107358" observedRunningTime="2025-12-11 08:54:56.332299821 +0000 UTC m=+1084.397715439" watchObservedRunningTime="2025-12-11 08:54:56.345700031 +0000 UTC m=+1084.411115639" Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.370517 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx" podStartSLOduration=6.129110843 podStartE2EDuration="50.370500071s" podCreationTimestamp="2025-12-11 08:54:06 +0000 UTC" firstStartedPulling="2025-12-11 08:54:10.770493653 +0000 UTC m=+1038.835909261" lastFinishedPulling="2025-12-11 08:54:55.011882881 +0000 UTC m=+1083.077298489" observedRunningTime="2025-12-11 08:54:56.365678329 +0000 UTC m=+1084.431093937" watchObservedRunningTime="2025-12-11 08:54:56.370500071 +0000 UTC m=+1084.435915679" Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.422008 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh" podStartSLOduration=5.071226785 podStartE2EDuration="49.421993937s" podCreationTimestamp="2025-12-11 08:54:07 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.627464552 +0000 UTC m=+1039.692880160" lastFinishedPulling="2025-12-11 08:54:55.978231694 +0000 UTC m=+1084.043647312" observedRunningTime="2025-12-11 08:54:56.420054976 +0000 UTC m=+1084.485470584" watchObservedRunningTime="2025-12-11 08:54:56.421993937 +0000 UTC m=+1084.487409545" Dec 11 08:54:56 crc kubenswrapper[4629]: I1211 08:54:56.422394 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9" podStartSLOduration=7.190408064 podStartE2EDuration="50.422388509s" podCreationTimestamp="2025-12-11 08:54:06 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.517721595 +0000 UTC m=+1039.583137203" lastFinishedPulling="2025-12-11 08:54:54.74970204 +0000 UTC m=+1082.815117648" observedRunningTime="2025-12-11 08:54:56.400836973 +0000 UTC m=+1084.466252591" watchObservedRunningTime="2025-12-11 08:54:56.422388509 +0000 UTC m=+1084.487804107" Dec 11 08:54:59 crc kubenswrapper[4629]: I1211 08:54:59.394480 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fr7wqh" Dec 11 08:55:00 crc kubenswrapper[4629]: I1211 08:55:00.267277 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-f8m45" Dec 11 08:55:03 crc kubenswrapper[4629]: I1211 08:55:03.058339 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-l59gh" Dec 11 08:55:07 crc kubenswrapper[4629]: I1211 08:55:07.003672 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p8lh9" Dec 11 08:55:07 crc kubenswrapper[4629]: I1211 08:55:07.013530 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-mhqnx" Dec 11 08:55:07 crc kubenswrapper[4629]: I1211 08:55:07.133765 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xdp6j" Dec 11 08:55:07 crc kubenswrapper[4629]: I1211 08:55:07.172703 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mtlt" Dec 11 08:55:07 crc kubenswrapper[4629]: I1211 08:55:07.536726 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-qxsl5" Dec 11 08:55:07 crc kubenswrapper[4629]: I1211 08:55:07.941432 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-p9wjh" Dec 11 08:55:09 crc kubenswrapper[4629]: I1211 08:55:09.486513 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" event={"ID":"94ec218d-3b08-463d-8225-f12cbcf0eb8d","Type":"ContainerStarted","Data":"437903c20fd7ee9e18ee42dfb2eacecfa9c40d698e5cb355c3fc7147a758129a"} Dec 11 08:55:09 crc kubenswrapper[4629]: I1211 08:55:09.488061 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" Dec 11 08:55:09 crc kubenswrapper[4629]: I1211 08:55:09.508796 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" podStartSLOduration=6.503031823 podStartE2EDuration="1m3.50877286s" podCreationTimestamp="2025-12-11 08:54:06 +0000 UTC" firstStartedPulling="2025-12-11 08:54:11.794249658 +0000 UTC m=+1039.859665266" lastFinishedPulling="2025-12-11 08:55:08.799990695 +0000 UTC m=+1096.865406303" observedRunningTime="2025-12-11 08:55:09.505892209 +0000 UTC m=+1097.571307817" watchObservedRunningTime="2025-12-11 08:55:09.50877286 +0000 UTC m=+1097.574188468" Dec 11 08:55:17 crc kubenswrapper[4629]: I1211 08:55:17.794048 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kqcvw" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.390402 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fjhv2"] Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.393109 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.397211 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-9n8rp" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.397486 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.397674 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.402153 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.403698 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fjhv2"] Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.467475 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xkhnp"] Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.469079 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.475133 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.483533 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xkhnp"] Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.513194 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8583b58d-bb36-4e76-8ccb-e6352330009f-config\") pod \"dnsmasq-dns-78dd6ddcc-xkhnp\" (UID: \"8583b58d-bb36-4e76-8ccb-e6352330009f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.513246 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbhsn\" (UniqueName: \"kubernetes.io/projected/d5bec238-d109-45e5-b2e1-c8e02f5b0710-kube-api-access-gbhsn\") pod \"dnsmasq-dns-675f4bcbfc-fjhv2\" (UID: \"d5bec238-d109-45e5-b2e1-c8e02f5b0710\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.513300 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5bec238-d109-45e5-b2e1-c8e02f5b0710-config\") pod \"dnsmasq-dns-675f4bcbfc-fjhv2\" (UID: \"d5bec238-d109-45e5-b2e1-c8e02f5b0710\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.513323 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z22xt\" (UniqueName: \"kubernetes.io/projected/8583b58d-bb36-4e76-8ccb-e6352330009f-kube-api-access-z22xt\") pod \"dnsmasq-dns-78dd6ddcc-xkhnp\" (UID: \"8583b58d-bb36-4e76-8ccb-e6352330009f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.513380 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8583b58d-bb36-4e76-8ccb-e6352330009f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-xkhnp\" (UID: \"8583b58d-bb36-4e76-8ccb-e6352330009f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.614513 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5bec238-d109-45e5-b2e1-c8e02f5b0710-config\") pod \"dnsmasq-dns-675f4bcbfc-fjhv2\" (UID: \"d5bec238-d109-45e5-b2e1-c8e02f5b0710\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.614586 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z22xt\" (UniqueName: \"kubernetes.io/projected/8583b58d-bb36-4e76-8ccb-e6352330009f-kube-api-access-z22xt\") pod \"dnsmasq-dns-78dd6ddcc-xkhnp\" (UID: \"8583b58d-bb36-4e76-8ccb-e6352330009f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.614633 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8583b58d-bb36-4e76-8ccb-e6352330009f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-xkhnp\" (UID: \"8583b58d-bb36-4e76-8ccb-e6352330009f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.614716 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8583b58d-bb36-4e76-8ccb-e6352330009f-config\") pod \"dnsmasq-dns-78dd6ddcc-xkhnp\" (UID: \"8583b58d-bb36-4e76-8ccb-e6352330009f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.614744 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbhsn\" (UniqueName: \"kubernetes.io/projected/d5bec238-d109-45e5-b2e1-c8e02f5b0710-kube-api-access-gbhsn\") pod \"dnsmasq-dns-675f4bcbfc-fjhv2\" (UID: \"d5bec238-d109-45e5-b2e1-c8e02f5b0710\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.615759 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8583b58d-bb36-4e76-8ccb-e6352330009f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-xkhnp\" (UID: \"8583b58d-bb36-4e76-8ccb-e6352330009f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.615903 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8583b58d-bb36-4e76-8ccb-e6352330009f-config\") pod \"dnsmasq-dns-78dd6ddcc-xkhnp\" (UID: \"8583b58d-bb36-4e76-8ccb-e6352330009f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.616674 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5bec238-d109-45e5-b2e1-c8e02f5b0710-config\") pod \"dnsmasq-dns-675f4bcbfc-fjhv2\" (UID: \"d5bec238-d109-45e5-b2e1-c8e02f5b0710\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.641802 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbhsn\" (UniqueName: \"kubernetes.io/projected/d5bec238-d109-45e5-b2e1-c8e02f5b0710-kube-api-access-gbhsn\") pod \"dnsmasq-dns-675f4bcbfc-fjhv2\" (UID: \"d5bec238-d109-45e5-b2e1-c8e02f5b0710\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.645188 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z22xt\" (UniqueName: \"kubernetes.io/projected/8583b58d-bb36-4e76-8ccb-e6352330009f-kube-api-access-z22xt\") pod \"dnsmasq-dns-78dd6ddcc-xkhnp\" (UID: \"8583b58d-bb36-4e76-8ccb-e6352330009f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.716628 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" Dec 11 08:55:32 crc kubenswrapper[4629]: I1211 08:55:32.787806 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:55:33 crc kubenswrapper[4629]: I1211 08:55:33.173393 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fjhv2"] Dec 11 08:55:33 crc kubenswrapper[4629]: I1211 08:55:33.275090 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xkhnp"] Dec 11 08:55:33 crc kubenswrapper[4629]: W1211 08:55:33.277403 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8583b58d_bb36_4e76_8ccb_e6352330009f.slice/crio-ff3f425f2aa877a835cbef30904d91b2a5f6472ba1ec687189c785561e31b070 WatchSource:0}: Error finding container ff3f425f2aa877a835cbef30904d91b2a5f6472ba1ec687189c785561e31b070: Status 404 returned error can't find the container with id ff3f425f2aa877a835cbef30904d91b2a5f6472ba1ec687189c785561e31b070 Dec 11 08:55:33 crc kubenswrapper[4629]: I1211 08:55:33.651238 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" event={"ID":"8583b58d-bb36-4e76-8ccb-e6352330009f","Type":"ContainerStarted","Data":"ff3f425f2aa877a835cbef30904d91b2a5f6472ba1ec687189c785561e31b070"} Dec 11 08:55:33 crc kubenswrapper[4629]: I1211 08:55:33.652097 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" event={"ID":"d5bec238-d109-45e5-b2e1-c8e02f5b0710","Type":"ContainerStarted","Data":"59b34c08ce63886fa497aa4bd633f427b9dc74aae6a3cc22f8b609a930682846"} Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.604624 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fjhv2"] Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.644892 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dxs2v"] Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.646579 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.658733 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dxs2v"] Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.764392 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b501db9-e5f4-484c-beb6-0085345af291-config\") pod \"dnsmasq-dns-666b6646f7-dxs2v\" (UID: \"6b501db9-e5f4-484c-beb6-0085345af291\") " pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.764504 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b501db9-e5f4-484c-beb6-0085345af291-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dxs2v\" (UID: \"6b501db9-e5f4-484c-beb6-0085345af291\") " pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.764530 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4ntq\" (UniqueName: \"kubernetes.io/projected/6b501db9-e5f4-484c-beb6-0085345af291-kube-api-access-p4ntq\") pod \"dnsmasq-dns-666b6646f7-dxs2v\" (UID: \"6b501db9-e5f4-484c-beb6-0085345af291\") " pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.866368 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b501db9-e5f4-484c-beb6-0085345af291-config\") pod \"dnsmasq-dns-666b6646f7-dxs2v\" (UID: \"6b501db9-e5f4-484c-beb6-0085345af291\") " pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.866436 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b501db9-e5f4-484c-beb6-0085345af291-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dxs2v\" (UID: \"6b501db9-e5f4-484c-beb6-0085345af291\") " pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.866466 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4ntq\" (UniqueName: \"kubernetes.io/projected/6b501db9-e5f4-484c-beb6-0085345af291-kube-api-access-p4ntq\") pod \"dnsmasq-dns-666b6646f7-dxs2v\" (UID: \"6b501db9-e5f4-484c-beb6-0085345af291\") " pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.868111 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b501db9-e5f4-484c-beb6-0085345af291-config\") pod \"dnsmasq-dns-666b6646f7-dxs2v\" (UID: \"6b501db9-e5f4-484c-beb6-0085345af291\") " pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.868948 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b501db9-e5f4-484c-beb6-0085345af291-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dxs2v\" (UID: \"6b501db9-e5f4-484c-beb6-0085345af291\") " pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.895080 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4ntq\" (UniqueName: \"kubernetes.io/projected/6b501db9-e5f4-484c-beb6-0085345af291-kube-api-access-p4ntq\") pod \"dnsmasq-dns-666b6646f7-dxs2v\" (UID: \"6b501db9-e5f4-484c-beb6-0085345af291\") " pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:55:35 crc kubenswrapper[4629]: I1211 08:55:35.987191 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.082572 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xkhnp"] Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.149985 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-hpgr4"] Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.151188 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.168303 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-hpgr4"] Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.281132 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwl2g\" (UniqueName: \"kubernetes.io/projected/b7a1e678-e517-4b2b-9915-323590b80ff5-kube-api-access-zwl2g\") pod \"dnsmasq-dns-57d769cc4f-hpgr4\" (UID: \"b7a1e678-e517-4b2b-9915-323590b80ff5\") " pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.281212 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a1e678-e517-4b2b-9915-323590b80ff5-config\") pod \"dnsmasq-dns-57d769cc4f-hpgr4\" (UID: \"b7a1e678-e517-4b2b-9915-323590b80ff5\") " pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.281314 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7a1e678-e517-4b2b-9915-323590b80ff5-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-hpgr4\" (UID: \"b7a1e678-e517-4b2b-9915-323590b80ff5\") " pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.382810 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a1e678-e517-4b2b-9915-323590b80ff5-config\") pod \"dnsmasq-dns-57d769cc4f-hpgr4\" (UID: \"b7a1e678-e517-4b2b-9915-323590b80ff5\") " pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.383241 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7a1e678-e517-4b2b-9915-323590b80ff5-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-hpgr4\" (UID: \"b7a1e678-e517-4b2b-9915-323590b80ff5\") " pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.383271 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwl2g\" (UniqueName: \"kubernetes.io/projected/b7a1e678-e517-4b2b-9915-323590b80ff5-kube-api-access-zwl2g\") pod \"dnsmasq-dns-57d769cc4f-hpgr4\" (UID: \"b7a1e678-e517-4b2b-9915-323590b80ff5\") " pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.384160 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7a1e678-e517-4b2b-9915-323590b80ff5-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-hpgr4\" (UID: \"b7a1e678-e517-4b2b-9915-323590b80ff5\") " pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.384683 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a1e678-e517-4b2b-9915-323590b80ff5-config\") pod \"dnsmasq-dns-57d769cc4f-hpgr4\" (UID: \"b7a1e678-e517-4b2b-9915-323590b80ff5\") " pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.408125 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwl2g\" (UniqueName: \"kubernetes.io/projected/b7a1e678-e517-4b2b-9915-323590b80ff5-kube-api-access-zwl2g\") pod \"dnsmasq-dns-57d769cc4f-hpgr4\" (UID: \"b7a1e678-e517-4b2b-9915-323590b80ff5\") " pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.489500 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.796383 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dxs2v"] Dec 11 08:55:36 crc kubenswrapper[4629]: W1211 08:55:36.811609 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b501db9_e5f4_484c_beb6_0085345af291.slice/crio-fb3ebdf3ff00c44f115c26bb21a54ce54643a8060fb4a7518bcd4d7a8dd0b185 WatchSource:0}: Error finding container fb3ebdf3ff00c44f115c26bb21a54ce54643a8060fb4a7518bcd4d7a8dd0b185: Status 404 returned error can't find the container with id fb3ebdf3ff00c44f115c26bb21a54ce54643a8060fb4a7518bcd4d7a8dd0b185 Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.874395 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.885921 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.887278 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.887862 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.889905 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.890049 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.890817 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.894055 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-t52c7" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.896064 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 11 08:55:36 crc kubenswrapper[4629]: I1211 08:55:36.906314 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.001584 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1a670dab-0f15-4572-b2a8-36ec2330adb2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.002152 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.002194 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.002247 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1a670dab-0f15-4572-b2a8-36ec2330adb2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.002271 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jddpj\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-kube-api-access-jddpj\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.002299 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.002329 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.002367 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.002386 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.002416 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-config-data\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.002449 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.011264 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-hpgr4"] Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.199142 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1a670dab-0f15-4572-b2a8-36ec2330adb2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.199189 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jddpj\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-kube-api-access-jddpj\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.199214 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.199235 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.199281 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.199298 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.199320 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-config-data\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.199362 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.199391 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1a670dab-0f15-4572-b2a8-36ec2330adb2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.199466 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.199508 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.201297 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.202830 4629 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.204426 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-config-data\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.204642 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.205269 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.207524 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.210379 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.211154 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.213059 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1a670dab-0f15-4572-b2a8-36ec2330adb2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.218652 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jddpj\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-kube-api-access-jddpj\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.236301 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1a670dab-0f15-4572-b2a8-36ec2330adb2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.250087 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.292866 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.294894 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.296725 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.296997 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.302482 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qgg55" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.302553 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.302707 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.302731 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.303503 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.317468 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.404654 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.404721 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.404947 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.405049 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.405150 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c633337-5e1d-438b-8b18-0acb62ec2da0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.405193 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.405214 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.405297 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c633337-5e1d-438b-8b18-0acb62ec2da0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.405333 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.405391 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.405458 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9zzh\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-kube-api-access-z9zzh\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.506892 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9zzh\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-kube-api-access-z9zzh\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.506955 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.507008 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.507036 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.507071 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.507110 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c633337-5e1d-438b-8b18-0acb62ec2da0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.507128 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.507147 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.507171 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c633337-5e1d-438b-8b18-0acb62ec2da0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.507187 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.507211 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.508807 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.508897 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.509653 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.509934 4629 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.512993 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c633337-5e1d-438b-8b18-0acb62ec2da0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.513751 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.514134 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.514409 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.519375 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c633337-5e1d-438b-8b18-0acb62ec2da0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.526050 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.541112 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.585050 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.604866 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9zzh\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-kube-api-access-z9zzh\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.667916 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.746154 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" event={"ID":"b7a1e678-e517-4b2b-9915-323590b80ff5","Type":"ContainerStarted","Data":"7ff762e75120b3985c26704b3698ac8492da02a4e3ce5ce5d3f609893a56d572"} Dec 11 08:55:37 crc kubenswrapper[4629]: I1211 08:55:37.762641 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" event={"ID":"6b501db9-e5f4-484c-beb6-0085345af291","Type":"ContainerStarted","Data":"fb3ebdf3ff00c44f115c26bb21a54ce54643a8060fb4a7518bcd4d7a8dd0b185"} Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.172681 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 08:55:38 crc kubenswrapper[4629]: W1211 08:55:38.190106 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a670dab_0f15_4572_b2a8_36ec2330adb2.slice/crio-7e87e1dc0f3024f6b70832ef4d267cce03220feb662ca7da8af51d7557f0ad31 WatchSource:0}: Error finding container 7e87e1dc0f3024f6b70832ef4d267cce03220feb662ca7da8af51d7557f0ad31: Status 404 returned error can't find the container with id 7e87e1dc0f3024f6b70832ef4d267cce03220feb662ca7da8af51d7557f0ad31 Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.331816 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 08:55:38 crc kubenswrapper[4629]: W1211 08:55:38.361504 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c633337_5e1d_438b_8b18_0acb62ec2da0.slice/crio-2b8d148c377e4e26f06811de5ae1ce5d87f41bc23dc8e83dbbf889b9a4847951 WatchSource:0}: Error finding container 2b8d148c377e4e26f06811de5ae1ce5d87f41bc23dc8e83dbbf889b9a4847951: Status 404 returned error can't find the container with id 2b8d148c377e4e26f06811de5ae1ce5d87f41bc23dc8e83dbbf889b9a4847951 Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.535893 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.537379 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.542311 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.542558 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.542933 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-f628n" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.548262 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.567777 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.574429 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.676264 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/762352f0-3934-4d44-b846-fa922ffedd1d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.676311 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762352f0-3934-4d44-b846-fa922ffedd1d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.676381 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/762352f0-3934-4d44-b846-fa922ffedd1d-kolla-config\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.676416 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.676534 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/762352f0-3934-4d44-b846-fa922ffedd1d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.676591 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/762352f0-3934-4d44-b846-fa922ffedd1d-config-data-default\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.676679 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgg6s\" (UniqueName: \"kubernetes.io/projected/762352f0-3934-4d44-b846-fa922ffedd1d-kube-api-access-bgg6s\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.676715 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762352f0-3934-4d44-b846-fa922ffedd1d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.777488 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762352f0-3934-4d44-b846-fa922ffedd1d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.777541 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/762352f0-3934-4d44-b846-fa922ffedd1d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.777563 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762352f0-3934-4d44-b846-fa922ffedd1d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.777626 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/762352f0-3934-4d44-b846-fa922ffedd1d-kolla-config\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.777659 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.777679 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/762352f0-3934-4d44-b846-fa922ffedd1d-config-data-default\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.777696 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/762352f0-3934-4d44-b846-fa922ffedd1d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.777721 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgg6s\" (UniqueName: \"kubernetes.io/projected/762352f0-3934-4d44-b846-fa922ffedd1d-kube-api-access-bgg6s\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.781448 4629 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.782884 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/762352f0-3934-4d44-b846-fa922ffedd1d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.784025 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762352f0-3934-4d44-b846-fa922ffedd1d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.785428 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/762352f0-3934-4d44-b846-fa922ffedd1d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.785649 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/762352f0-3934-4d44-b846-fa922ffedd1d-config-data-default\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.787094 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/762352f0-3934-4d44-b846-fa922ffedd1d-kolla-config\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.787974 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1a670dab-0f15-4572-b2a8-36ec2330adb2","Type":"ContainerStarted","Data":"7e87e1dc0f3024f6b70832ef4d267cce03220feb662ca7da8af51d7557f0ad31"} Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.788780 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0c633337-5e1d-438b-8b18-0acb62ec2da0","Type":"ContainerStarted","Data":"2b8d148c377e4e26f06811de5ae1ce5d87f41bc23dc8e83dbbf889b9a4847951"} Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.819754 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgg6s\" (UniqueName: \"kubernetes.io/projected/762352f0-3934-4d44-b846-fa922ffedd1d-kube-api-access-bgg6s\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.830533 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762352f0-3934-4d44-b846-fa922ffedd1d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.865287 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"762352f0-3934-4d44-b846-fa922ffedd1d\") " pod="openstack/openstack-galera-0" Dec 11 08:55:38 crc kubenswrapper[4629]: I1211 08:55:38.871531 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.740272 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.743003 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.746800 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.749272 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.752937 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.753026 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-q2t7l" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.753151 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.896422 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ab910fe2-4e06-4174-9638-e9b920c5bff9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.896580 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab910fe2-4e06-4174-9638-e9b920c5bff9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.896649 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab910fe2-4e06-4174-9638-e9b920c5bff9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.896744 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab910fe2-4e06-4174-9638-e9b920c5bff9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.896817 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.896839 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ab910fe2-4e06-4174-9638-e9b920c5bff9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.896896 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ab910fe2-4e06-4174-9638-e9b920c5bff9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.896987 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjbrx\" (UniqueName: \"kubernetes.io/projected/ab910fe2-4e06-4174-9638-e9b920c5bff9-kube-api-access-fjbrx\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.998490 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ab910fe2-4e06-4174-9638-e9b920c5bff9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.998578 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab910fe2-4e06-4174-9638-e9b920c5bff9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.998609 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab910fe2-4e06-4174-9638-e9b920c5bff9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.998661 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab910fe2-4e06-4174-9638-e9b920c5bff9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.998706 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.998727 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ab910fe2-4e06-4174-9638-e9b920c5bff9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.998750 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ab910fe2-4e06-4174-9638-e9b920c5bff9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:39 crc kubenswrapper[4629]: I1211 08:55:39.998768 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjbrx\" (UniqueName: \"kubernetes.io/projected/ab910fe2-4e06-4174-9638-e9b920c5bff9-kube-api-access-fjbrx\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.000452 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ab910fe2-4e06-4174-9638-e9b920c5bff9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.000642 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ab910fe2-4e06-4174-9638-e9b920c5bff9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.000878 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ab910fe2-4e06-4174-9638-e9b920c5bff9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.001279 4629 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.002129 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab910fe2-4e06-4174-9638-e9b920c5bff9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.013767 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab910fe2-4e06-4174-9638-e9b920c5bff9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.030461 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab910fe2-4e06-4174-9638-e9b920c5bff9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.030975 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjbrx\" (UniqueName: \"kubernetes.io/projected/ab910fe2-4e06-4174-9638-e9b920c5bff9-kube-api-access-fjbrx\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.043973 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ab910fe2-4e06-4174-9638-e9b920c5bff9\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.084188 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.577445 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.592776 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.610427 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.610743 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-hccqt" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.638324 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.652821 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.740516 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-config-data\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.740656 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.740714 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.741420 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-kolla-config\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.741475 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjhbq\" (UniqueName: \"kubernetes.io/projected/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-kube-api-access-hjhbq\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.844993 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-kolla-config\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.845051 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjhbq\" (UniqueName: \"kubernetes.io/projected/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-kube-api-access-hjhbq\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.845132 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-config-data\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.845174 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.845209 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.846303 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-config-data\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.848338 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-kolla-config\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.863494 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.866613 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjhbq\" (UniqueName: \"kubernetes.io/projected/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-kube-api-access-hjhbq\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.868442 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/147dc374-3f14-40f8-bf2f-ab1bd5a72d4e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e\") " pod="openstack/memcached-0" Dec 11 08:55:40 crc kubenswrapper[4629]: I1211 08:55:40.970353 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 11 08:55:42 crc kubenswrapper[4629]: I1211 08:55:42.398023 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 08:55:42 crc kubenswrapper[4629]: I1211 08:55:42.399759 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 08:55:42 crc kubenswrapper[4629]: I1211 08:55:42.402118 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-rsnzq" Dec 11 08:55:42 crc kubenswrapper[4629]: I1211 08:55:42.425374 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 08:55:42 crc kubenswrapper[4629]: I1211 08:55:42.673504 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8jnv\" (UniqueName: \"kubernetes.io/projected/2594a86c-caf2-4874-a415-4e6fda88db08-kube-api-access-d8jnv\") pod \"kube-state-metrics-0\" (UID: \"2594a86c-caf2-4874-a415-4e6fda88db08\") " pod="openstack/kube-state-metrics-0" Dec 11 08:55:42 crc kubenswrapper[4629]: I1211 08:55:42.774470 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8jnv\" (UniqueName: \"kubernetes.io/projected/2594a86c-caf2-4874-a415-4e6fda88db08-kube-api-access-d8jnv\") pod \"kube-state-metrics-0\" (UID: \"2594a86c-caf2-4874-a415-4e6fda88db08\") " pod="openstack/kube-state-metrics-0" Dec 11 08:55:42 crc kubenswrapper[4629]: I1211 08:55:42.792607 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8jnv\" (UniqueName: \"kubernetes.io/projected/2594a86c-caf2-4874-a415-4e6fda88db08-kube-api-access-d8jnv\") pod \"kube-state-metrics-0\" (UID: \"2594a86c-caf2-4874-a415-4e6fda88db08\") " pod="openstack/kube-state-metrics-0" Dec 11 08:55:43 crc kubenswrapper[4629]: I1211 08:55:43.007523 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.779602 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.781548 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.785783 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.786325 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-xrklt" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.786440 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.790943 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.791137 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.802766 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.961742 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0ccfb64c-586b-4381-96b7-5846746d6433-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.961789 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ccfb64c-586b-4381-96b7-5846746d6433-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.961828 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ccfb64c-586b-4381-96b7-5846746d6433-config\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.961866 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccfb64c-586b-4381-96b7-5846746d6433-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.962110 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ccfb64c-586b-4381-96b7-5846746d6433-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.962299 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.962350 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ccfb64c-586b-4381-96b7-5846746d6433-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:45 crc kubenswrapper[4629]: I1211 08:55:45.962415 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5wnz\" (UniqueName: \"kubernetes.io/projected/0ccfb64c-586b-4381-96b7-5846746d6433-kube-api-access-w5wnz\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.064642 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ccfb64c-586b-4381-96b7-5846746d6433-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.065989 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.066100 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ccfb64c-586b-4381-96b7-5846746d6433-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.066189 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5wnz\" (UniqueName: \"kubernetes.io/projected/0ccfb64c-586b-4381-96b7-5846746d6433-kube-api-access-w5wnz\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.066355 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0ccfb64c-586b-4381-96b7-5846746d6433-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.066467 4629 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.066808 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0ccfb64c-586b-4381-96b7-5846746d6433-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.066493 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ccfb64c-586b-4381-96b7-5846746d6433-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.067035 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ccfb64c-586b-4381-96b7-5846746d6433-config\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.067108 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccfb64c-586b-4381-96b7-5846746d6433-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.067644 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ccfb64c-586b-4381-96b7-5846746d6433-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.067792 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ccfb64c-586b-4381-96b7-5846746d6433-config\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.082753 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ccfb64c-586b-4381-96b7-5846746d6433-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.086479 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ccfb64c-586b-4381-96b7-5846746d6433-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.086525 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5wnz\" (UniqueName: \"kubernetes.io/projected/0ccfb64c-586b-4381-96b7-5846746d6433-kube-api-access-w5wnz\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.090667 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccfb64c-586b-4381-96b7-5846746d6433-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.098214 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"0ccfb64c-586b-4381-96b7-5846746d6433\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.105508 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.941094 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8lfjz"] Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.942438 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.947653 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.948261 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-rrvcb" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.948559 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 11 08:55:46 crc kubenswrapper[4629]: I1211 08:55:46.959838 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8lfjz"] Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.032509 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-v9ctv"] Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.035519 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.088091 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-scripts\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.088533 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/708ce743-d291-41e4-8c15-a9ad1cda240e-var-log-ovn\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.088557 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-etc-ovs\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.088575 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/708ce743-d291-41e4-8c15-a9ad1cda240e-var-run-ovn\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.088614 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b72n5\" (UniqueName: \"kubernetes.io/projected/708ce743-d291-41e4-8c15-a9ad1cda240e-kube-api-access-b72n5\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.088636 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/708ce743-d291-41e4-8c15-a9ad1cda240e-ovn-controller-tls-certs\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.088693 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-var-run\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.088711 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/708ce743-d291-41e4-8c15-a9ad1cda240e-combined-ca-bundle\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.088793 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbr9m\" (UniqueName: \"kubernetes.io/projected/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-kube-api-access-mbr9m\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.088891 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-var-lib\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.088930 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-var-log\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.088965 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/708ce743-d291-41e4-8c15-a9ad1cda240e-var-run\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.089046 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/708ce743-d291-41e4-8c15-a9ad1cda240e-scripts\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.089873 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-v9ctv"] Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.189597 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-etc-ovs\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.189646 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/708ce743-d291-41e4-8c15-a9ad1cda240e-var-run-ovn\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.189675 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b72n5\" (UniqueName: \"kubernetes.io/projected/708ce743-d291-41e4-8c15-a9ad1cda240e-kube-api-access-b72n5\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.189696 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/708ce743-d291-41e4-8c15-a9ad1cda240e-ovn-controller-tls-certs\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.189733 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-var-run\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.189757 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/708ce743-d291-41e4-8c15-a9ad1cda240e-combined-ca-bundle\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.189785 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbr9m\" (UniqueName: \"kubernetes.io/projected/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-kube-api-access-mbr9m\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.189831 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-var-lib\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.189994 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-var-log\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.190024 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/708ce743-d291-41e4-8c15-a9ad1cda240e-var-run\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.190065 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/708ce743-d291-41e4-8c15-a9ad1cda240e-scripts\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.190090 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-scripts\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.190127 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/708ce743-d291-41e4-8c15-a9ad1cda240e-var-log-ovn\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.190476 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-etc-ovs\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.190591 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/708ce743-d291-41e4-8c15-a9ad1cda240e-var-log-ovn\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.190647 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/708ce743-d291-41e4-8c15-a9ad1cda240e-var-run-ovn\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.190755 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-var-lib\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.190825 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-var-run\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.191175 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/708ce743-d291-41e4-8c15-a9ad1cda240e-var-run\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.191277 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-var-log\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.193776 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/708ce743-d291-41e4-8c15-a9ad1cda240e-scripts\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.206827 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-scripts\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.213497 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/708ce743-d291-41e4-8c15-a9ad1cda240e-combined-ca-bundle\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.217415 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/708ce743-d291-41e4-8c15-a9ad1cda240e-ovn-controller-tls-certs\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.217654 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b72n5\" (UniqueName: \"kubernetes.io/projected/708ce743-d291-41e4-8c15-a9ad1cda240e-kube-api-access-b72n5\") pod \"ovn-controller-8lfjz\" (UID: \"708ce743-d291-41e4-8c15-a9ad1cda240e\") " pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.223171 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbr9m\" (UniqueName: \"kubernetes.io/projected/5a52781a-0193-4e3a-b0d3-4ee50d507f1a-kube-api-access-mbr9m\") pod \"ovn-controller-ovs-v9ctv\" (UID: \"5a52781a-0193-4e3a-b0d3-4ee50d507f1a\") " pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.264770 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8lfjz" Dec 11 08:55:47 crc kubenswrapper[4629]: I1211 08:55:47.358741 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.780395 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.783283 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.785899 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.786213 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.786704 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-zrpqg" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.787372 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.806350 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.936463 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.936517 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg4vt\" (UniqueName: \"kubernetes.io/projected/d2811a7c-5ac9-4db5-81d3-61771795a87e-kube-api-access-gg4vt\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.936554 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2811a7c-5ac9-4db5-81d3-61771795a87e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.936578 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d2811a7c-5ac9-4db5-81d3-61771795a87e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.936606 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2811a7c-5ac9-4db5-81d3-61771795a87e-config\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.936677 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2811a7c-5ac9-4db5-81d3-61771795a87e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.936700 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2811a7c-5ac9-4db5-81d3-61771795a87e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:49 crc kubenswrapper[4629]: I1211 08:55:49.936727 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2811a7c-5ac9-4db5-81d3-61771795a87e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.038011 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg4vt\" (UniqueName: \"kubernetes.io/projected/d2811a7c-5ac9-4db5-81d3-61771795a87e-kube-api-access-gg4vt\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.038077 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2811a7c-5ac9-4db5-81d3-61771795a87e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.038097 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d2811a7c-5ac9-4db5-81d3-61771795a87e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.038114 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2811a7c-5ac9-4db5-81d3-61771795a87e-config\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.038171 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2811a7c-5ac9-4db5-81d3-61771795a87e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.038191 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2811a7c-5ac9-4db5-81d3-61771795a87e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.038217 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2811a7c-5ac9-4db5-81d3-61771795a87e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.039121 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d2811a7c-5ac9-4db5-81d3-61771795a87e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.039498 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2811a7c-5ac9-4db5-81d3-61771795a87e-config\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.039994 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2811a7c-5ac9-4db5-81d3-61771795a87e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.043545 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2811a7c-5ac9-4db5-81d3-61771795a87e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.043766 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2811a7c-5ac9-4db5-81d3-61771795a87e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.054839 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2811a7c-5ac9-4db5-81d3-61771795a87e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.058534 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg4vt\" (UniqueName: \"kubernetes.io/projected/d2811a7c-5ac9-4db5-81d3-61771795a87e-kube-api-access-gg4vt\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.139893 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.140326 4629 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.158800 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d2811a7c-5ac9-4db5-81d3-61771795a87e\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.411477 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.486517 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:55:50 crc kubenswrapper[4629]: I1211 08:55:50.486916 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:55:58 crc kubenswrapper[4629]: I1211 08:55:58.253012 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 11 08:55:58 crc kubenswrapper[4629]: E1211 08:55:58.743914 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 11 08:55:58 crc kubenswrapper[4629]: E1211 08:55:58.744403 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z22xt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-xkhnp_openstack(8583b58d-bb36-4e76-8ccb-e6352330009f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:55:58 crc kubenswrapper[4629]: E1211 08:55:58.745620 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" podUID="8583b58d-bb36-4e76-8ccb-e6352330009f" Dec 11 08:55:59 crc kubenswrapper[4629]: W1211 08:55:59.776089 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod147dc374_3f14_40f8_bf2f_ab1bd5a72d4e.slice/crio-ea83608bb1bc2ed4ea857cdb2377035f2a52e6beb8ea4e82a1156e022bef7e94 WatchSource:0}: Error finding container ea83608bb1bc2ed4ea857cdb2377035f2a52e6beb8ea4e82a1156e022bef7e94: Status 404 returned error can't find the container with id ea83608bb1bc2ed4ea857cdb2377035f2a52e6beb8ea4e82a1156e022bef7e94 Dec 11 08:55:59 crc kubenswrapper[4629]: E1211 08:55:59.819607 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 11 08:55:59 crc kubenswrapper[4629]: E1211 08:55:59.819795 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gbhsn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-fjhv2_openstack(d5bec238-d109-45e5-b2e1-c8e02f5b0710): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:55:59 crc kubenswrapper[4629]: E1211 08:55:59.822324 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" podUID="d5bec238-d109-45e5-b2e1-c8e02f5b0710" Dec 11 08:55:59 crc kubenswrapper[4629]: E1211 08:55:59.824976 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 11 08:55:59 crc kubenswrapper[4629]: E1211 08:55:59.825270 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zwl2g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-hpgr4_openstack(b7a1e678-e517-4b2b-9915-323590b80ff5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:55:59 crc kubenswrapper[4629]: E1211 08:55:59.829857 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" podUID="b7a1e678-e517-4b2b-9915-323590b80ff5" Dec 11 08:55:59 crc kubenswrapper[4629]: E1211 08:55:59.830042 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 11 08:55:59 crc kubenswrapper[4629]: E1211 08:55:59.830248 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p4ntq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-dxs2v_openstack(6b501db9-e5f4-484c-beb6-0085345af291): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:55:59 crc kubenswrapper[4629]: E1211 08:55:59.832151 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" podUID="6b501db9-e5f4-484c-beb6-0085345af291" Dec 11 08:55:59 crc kubenswrapper[4629]: I1211 08:55:59.966527 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.039434 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.039452 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-xkhnp" event={"ID":"8583b58d-bb36-4e76-8ccb-e6352330009f","Type":"ContainerDied","Data":"ff3f425f2aa877a835cbef30904d91b2a5f6472ba1ec687189c785561e31b070"} Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.048194 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e","Type":"ContainerStarted","Data":"ea83608bb1bc2ed4ea857cdb2377035f2a52e6beb8ea4e82a1156e022bef7e94"} Dec 11 08:56:00 crc kubenswrapper[4629]: E1211 08:56:00.073924 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" podUID="6b501db9-e5f4-484c-beb6-0085345af291" Dec 11 08:56:00 crc kubenswrapper[4629]: E1211 08:56:00.074146 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" podUID="b7a1e678-e517-4b2b-9915-323590b80ff5" Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.157382 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8583b58d-bb36-4e76-8ccb-e6352330009f-config\") pod \"8583b58d-bb36-4e76-8ccb-e6352330009f\" (UID: \"8583b58d-bb36-4e76-8ccb-e6352330009f\") " Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.157430 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z22xt\" (UniqueName: \"kubernetes.io/projected/8583b58d-bb36-4e76-8ccb-e6352330009f-kube-api-access-z22xt\") pod \"8583b58d-bb36-4e76-8ccb-e6352330009f\" (UID: \"8583b58d-bb36-4e76-8ccb-e6352330009f\") " Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.157483 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8583b58d-bb36-4e76-8ccb-e6352330009f-dns-svc\") pod \"8583b58d-bb36-4e76-8ccb-e6352330009f\" (UID: \"8583b58d-bb36-4e76-8ccb-e6352330009f\") " Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.161298 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8583b58d-bb36-4e76-8ccb-e6352330009f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8583b58d-bb36-4e76-8ccb-e6352330009f" (UID: "8583b58d-bb36-4e76-8ccb-e6352330009f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.171588 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8583b58d-bb36-4e76-8ccb-e6352330009f-config" (OuterVolumeSpecName: "config") pod "8583b58d-bb36-4e76-8ccb-e6352330009f" (UID: "8583b58d-bb36-4e76-8ccb-e6352330009f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.177050 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8583b58d-bb36-4e76-8ccb-e6352330009f-kube-api-access-z22xt" (OuterVolumeSpecName: "kube-api-access-z22xt") pod "8583b58d-bb36-4e76-8ccb-e6352330009f" (UID: "8583b58d-bb36-4e76-8ccb-e6352330009f"). InnerVolumeSpecName "kube-api-access-z22xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.259365 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8583b58d-bb36-4e76-8ccb-e6352330009f-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.259402 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z22xt\" (UniqueName: \"kubernetes.io/projected/8583b58d-bb36-4e76-8ccb-e6352330009f-kube-api-access-z22xt\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.259418 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8583b58d-bb36-4e76-8ccb-e6352330009f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.398005 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xkhnp"] Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.417463 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xkhnp"] Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.456146 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.512167 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8lfjz"] Dec 11 08:56:00 crc kubenswrapper[4629]: W1211 08:56:00.518267 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod708ce743_d291_41e4_8c15_a9ad1cda240e.slice/crio-dca6eea00aa3b0c145cf57889ad85dfe18e11f817a3c3a8b90cdbec60029e70e WatchSource:0}: Error finding container dca6eea00aa3b0c145cf57889ad85dfe18e11f817a3c3a8b90cdbec60029e70e: Status 404 returned error can't find the container with id dca6eea00aa3b0c145cf57889ad85dfe18e11f817a3c3a8b90cdbec60029e70e Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.708248 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 08:56:00 crc kubenswrapper[4629]: W1211 08:56:00.739316 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2594a86c_caf2_4874_a415_4e6fda88db08.slice/crio-a8a55aaa6287412e36c6e0c6b921d76515e9b2c68eab039ac9479d99d9f4ac3d WatchSource:0}: Error finding container a8a55aaa6287412e36c6e0c6b921d76515e9b2c68eab039ac9479d99d9f4ac3d: Status 404 returned error can't find the container with id a8a55aaa6287412e36c6e0c6b921d76515e9b2c68eab039ac9479d99d9f4ac3d Dec 11 08:56:00 crc kubenswrapper[4629]: I1211 08:56:00.910510 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 11 08:56:00 crc kubenswrapper[4629]: W1211 08:56:00.967266 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod762352f0_3934_4d44_b846_fa922ffedd1d.slice/crio-b0c2a719f8426b4fcfe860fd043ed4bb8c2fe7e26111e634198fadd814cba6f3 WatchSource:0}: Error finding container b0c2a719f8426b4fcfe860fd043ed4bb8c2fe7e26111e634198fadd814cba6f3: Status 404 returned error can't find the container with id b0c2a719f8426b4fcfe860fd043ed4bb8c2fe7e26111e634198fadd814cba6f3 Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.010733 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-v9ctv"] Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.056397 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ab910fe2-4e06-4174-9638-e9b920c5bff9","Type":"ContainerStarted","Data":"2b0f790b7b5666849745563a01f0724a928a24766774e5d9a311685c1780ff08"} Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.060200 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"762352f0-3934-4d44-b846-fa922ffedd1d","Type":"ContainerStarted","Data":"b0c2a719f8426b4fcfe860fd043ed4bb8c2fe7e26111e634198fadd814cba6f3"} Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.061497 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8lfjz" event={"ID":"708ce743-d291-41e4-8c15-a9ad1cda240e","Type":"ContainerStarted","Data":"dca6eea00aa3b0c145cf57889ad85dfe18e11f817a3c3a8b90cdbec60029e70e"} Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.063693 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" event={"ID":"d5bec238-d109-45e5-b2e1-c8e02f5b0710","Type":"ContainerDied","Data":"59b34c08ce63886fa497aa4bd633f427b9dc74aae6a3cc22f8b609a930682846"} Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.063736 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59b34c08ce63886fa497aa4bd633f427b9dc74aae6a3cc22f8b609a930682846" Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.065701 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2594a86c-caf2-4874-a415-4e6fda88db08","Type":"ContainerStarted","Data":"a8a55aaa6287412e36c6e0c6b921d76515e9b2c68eab039ac9479d99d9f4ac3d"} Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.067016 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v9ctv" event={"ID":"5a52781a-0193-4e3a-b0d3-4ee50d507f1a","Type":"ContainerStarted","Data":"9c52a3d663f0873bf307f7c1f9252e31a3d21035f15a8a0f46032b603d070a14"} Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.121130 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.170106 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" Dec 11 08:56:01 crc kubenswrapper[4629]: W1211 08:56:01.236963 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2811a7c_5ac9_4db5_81d3_61771795a87e.slice/crio-d8a4b6757b22497f856e9f98f47575ff16fe91644e1bdbfddf645eb9abdc9add WatchSource:0}: Error finding container d8a4b6757b22497f856e9f98f47575ff16fe91644e1bdbfddf645eb9abdc9add: Status 404 returned error can't find the container with id d8a4b6757b22497f856e9f98f47575ff16fe91644e1bdbfddf645eb9abdc9add Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.291218 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5bec238-d109-45e5-b2e1-c8e02f5b0710-config\") pod \"d5bec238-d109-45e5-b2e1-c8e02f5b0710\" (UID: \"d5bec238-d109-45e5-b2e1-c8e02f5b0710\") " Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.291283 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbhsn\" (UniqueName: \"kubernetes.io/projected/d5bec238-d109-45e5-b2e1-c8e02f5b0710-kube-api-access-gbhsn\") pod \"d5bec238-d109-45e5-b2e1-c8e02f5b0710\" (UID: \"d5bec238-d109-45e5-b2e1-c8e02f5b0710\") " Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.292365 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5bec238-d109-45e5-b2e1-c8e02f5b0710-config" (OuterVolumeSpecName: "config") pod "d5bec238-d109-45e5-b2e1-c8e02f5b0710" (UID: "d5bec238-d109-45e5-b2e1-c8e02f5b0710"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.295818 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5bec238-d109-45e5-b2e1-c8e02f5b0710-kube-api-access-gbhsn" (OuterVolumeSpecName: "kube-api-access-gbhsn") pod "d5bec238-d109-45e5-b2e1-c8e02f5b0710" (UID: "d5bec238-d109-45e5-b2e1-c8e02f5b0710"). InnerVolumeSpecName "kube-api-access-gbhsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.394969 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5bec238-d109-45e5-b2e1-c8e02f5b0710-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:01 crc kubenswrapper[4629]: I1211 08:56:01.395050 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbhsn\" (UniqueName: \"kubernetes.io/projected/d5bec238-d109-45e5-b2e1-c8e02f5b0710-kube-api-access-gbhsn\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:02 crc kubenswrapper[4629]: I1211 08:56:02.084554 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0c633337-5e1d-438b-8b18-0acb62ec2da0","Type":"ContainerStarted","Data":"4a5a857a435fc978860280f4cbe148a7f4dd2ee9a3c7d2ab96099c938c4b272a"} Dec 11 08:56:02 crc kubenswrapper[4629]: I1211 08:56:02.088833 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 11 08:56:02 crc kubenswrapper[4629]: I1211 08:56:02.089085 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d2811a7c-5ac9-4db5-81d3-61771795a87e","Type":"ContainerStarted","Data":"d8a4b6757b22497f856e9f98f47575ff16fe91644e1bdbfddf645eb9abdc9add"} Dec 11 08:56:02 crc kubenswrapper[4629]: I1211 08:56:02.091030 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1a670dab-0f15-4572-b2a8-36ec2330adb2","Type":"ContainerStarted","Data":"d3ec5182b5c5afada44fd8ab808ed3ad84b1a4b2fbc5e3bd1a17034f8d925a9a"} Dec 11 08:56:02 crc kubenswrapper[4629]: I1211 08:56:02.091063 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fjhv2" Dec 11 08:56:02 crc kubenswrapper[4629]: I1211 08:56:02.192174 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fjhv2"] Dec 11 08:56:02 crc kubenswrapper[4629]: I1211 08:56:02.197241 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fjhv2"] Dec 11 08:56:02 crc kubenswrapper[4629]: I1211 08:56:02.221718 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8583b58d-bb36-4e76-8ccb-e6352330009f" path="/var/lib/kubelet/pods/8583b58d-bb36-4e76-8ccb-e6352330009f/volumes" Dec 11 08:56:02 crc kubenswrapper[4629]: I1211 08:56:02.222129 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5bec238-d109-45e5-b2e1-c8e02f5b0710" path="/var/lib/kubelet/pods/d5bec238-d109-45e5-b2e1-c8e02f5b0710/volumes" Dec 11 08:56:03 crc kubenswrapper[4629]: I1211 08:56:03.102080 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"0ccfb64c-586b-4381-96b7-5846746d6433","Type":"ContainerStarted","Data":"d602ca44eb66a92f740fcc013222cea916ef338945b199d82c84b171c12197f8"} Dec 11 08:56:06 crc kubenswrapper[4629]: I1211 08:56:06.127099 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"147dc374-3f14-40f8-bf2f-ab1bd5a72d4e","Type":"ContainerStarted","Data":"672098ad484b143c9a28d98ae496ab180a67e5607ec389cb9e3d7aa7c3874c61"} Dec 11 08:56:06 crc kubenswrapper[4629]: I1211 08:56:06.127533 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 11 08:56:06 crc kubenswrapper[4629]: I1211 08:56:06.155495 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=20.439475574 podStartE2EDuration="26.15546303s" podCreationTimestamp="2025-12-11 08:55:40 +0000 UTC" firstStartedPulling="2025-12-11 08:55:59.799049634 +0000 UTC m=+1147.864465242" lastFinishedPulling="2025-12-11 08:56:05.51503708 +0000 UTC m=+1153.580452698" observedRunningTime="2025-12-11 08:56:06.148603324 +0000 UTC m=+1154.214018922" watchObservedRunningTime="2025-12-11 08:56:06.15546303 +0000 UTC m=+1154.220878638" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.327072 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-vzhn9"] Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.333127 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.342682 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.353663 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vzhn9"] Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.517732 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dxs2v"] Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.518934 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-combined-ca-bundle\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.518979 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-config\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.519021 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-ovs-rundir\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.519065 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-ovn-rundir\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.519215 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjbzn\" (UniqueName: \"kubernetes.io/projected/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-kube-api-access-pjbzn\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.519275 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.585307 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-pf5fp"] Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.593047 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.626042 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.626447 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-pf5fp"] Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.632196 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-config\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.632273 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-ovs-rundir\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.632335 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-ovn-rundir\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.632372 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9fk9\" (UniqueName: \"kubernetes.io/projected/902107e5-7c95-4d45-b298-93b32294cfb6-kube-api-access-h9fk9\") pod \"dnsmasq-dns-6bc7876d45-pf5fp\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.632437 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjbzn\" (UniqueName: \"kubernetes.io/projected/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-kube-api-access-pjbzn\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.632477 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-pf5fp\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.632501 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-config\") pod \"dnsmasq-dns-6bc7876d45-pf5fp\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.632518 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.632594 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-pf5fp\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.632612 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-combined-ca-bundle\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.633096 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-config\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.633309 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-ovs-rundir\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.633327 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-ovn-rundir\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.643175 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-combined-ca-bundle\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.661501 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.670495 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjbzn\" (UniqueName: \"kubernetes.io/projected/8f807dbd-baed-4ee5-b422-8a3a4e2d045f-kube-api-access-pjbzn\") pod \"ovn-controller-metrics-vzhn9\" (UID: \"8f807dbd-baed-4ee5-b422-8a3a4e2d045f\") " pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.735529 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-pf5fp\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.735639 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9fk9\" (UniqueName: \"kubernetes.io/projected/902107e5-7c95-4d45-b298-93b32294cfb6-kube-api-access-h9fk9\") pod \"dnsmasq-dns-6bc7876d45-pf5fp\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.735703 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-pf5fp\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.735733 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-config\") pod \"dnsmasq-dns-6bc7876d45-pf5fp\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.736810 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-config\") pod \"dnsmasq-dns-6bc7876d45-pf5fp\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.737694 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-pf5fp\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.739493 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-pf5fp\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.761922 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9fk9\" (UniqueName: \"kubernetes.io/projected/902107e5-7c95-4d45-b298-93b32294cfb6-kube-api-access-h9fk9\") pod \"dnsmasq-dns-6bc7876d45-pf5fp\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.775131 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-hpgr4"] Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.869516 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-vdhgz"] Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.877982 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.887812 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.908808 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-vdhgz"] Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.968893 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vzhn9" Dec 11 08:56:10 crc kubenswrapper[4629]: I1211 08:56:10.979013 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.026402 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.044027 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.044098 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-dns-svc\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.044152 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.044173 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-config\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.044203 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frls5\" (UniqueName: \"kubernetes.io/projected/6ddc853e-11fc-4d38-b550-8b42d1997c2b-kube-api-access-frls5\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.146054 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-dns-svc\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.146173 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.146202 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-config\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.146241 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frls5\" (UniqueName: \"kubernetes.io/projected/6ddc853e-11fc-4d38-b550-8b42d1997c2b-kube-api-access-frls5\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.146347 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.149138 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-dns-svc\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.149553 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-config\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.150126 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.150408 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.168275 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frls5\" (UniqueName: \"kubernetes.io/projected/6ddc853e-11fc-4d38-b550-8b42d1997c2b-kube-api-access-frls5\") pod \"dnsmasq-dns-8554648995-vdhgz\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.175714 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" event={"ID":"6b501db9-e5f4-484c-beb6-0085345af291","Type":"ContainerDied","Data":"fb3ebdf3ff00c44f115c26bb21a54ce54643a8060fb4a7518bcd4d7a8dd0b185"} Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.175780 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb3ebdf3ff00c44f115c26bb21a54ce54643a8060fb4a7518bcd4d7a8dd0b185" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.175878 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.251681 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.348467 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4ntq\" (UniqueName: \"kubernetes.io/projected/6b501db9-e5f4-484c-beb6-0085345af291-kube-api-access-p4ntq\") pod \"6b501db9-e5f4-484c-beb6-0085345af291\" (UID: \"6b501db9-e5f4-484c-beb6-0085345af291\") " Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.348722 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b501db9-e5f4-484c-beb6-0085345af291-config\") pod \"6b501db9-e5f4-484c-beb6-0085345af291\" (UID: \"6b501db9-e5f4-484c-beb6-0085345af291\") " Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.348795 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b501db9-e5f4-484c-beb6-0085345af291-dns-svc\") pod \"6b501db9-e5f4-484c-beb6-0085345af291\" (UID: \"6b501db9-e5f4-484c-beb6-0085345af291\") " Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.349331 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b501db9-e5f4-484c-beb6-0085345af291-config" (OuterVolumeSpecName: "config") pod "6b501db9-e5f4-484c-beb6-0085345af291" (UID: "6b501db9-e5f4-484c-beb6-0085345af291"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.351029 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b501db9-e5f4-484c-beb6-0085345af291-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b501db9-e5f4-484c-beb6-0085345af291" (UID: "6b501db9-e5f4-484c-beb6-0085345af291"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.352754 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b501db9-e5f4-484c-beb6-0085345af291-kube-api-access-p4ntq" (OuterVolumeSpecName: "kube-api-access-p4ntq") pod "6b501db9-e5f4-484c-beb6-0085345af291" (UID: "6b501db9-e5f4-484c-beb6-0085345af291"). InnerVolumeSpecName "kube-api-access-p4ntq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.455938 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b501db9-e5f4-484c-beb6-0085345af291-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.456015 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b501db9-e5f4-484c-beb6-0085345af291-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.456027 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4ntq\" (UniqueName: \"kubernetes.io/projected/6b501db9-e5f4-484c-beb6-0085345af291-kube-api-access-p4ntq\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.497921 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.659534 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a1e678-e517-4b2b-9915-323590b80ff5-config\") pod \"b7a1e678-e517-4b2b-9915-323590b80ff5\" (UID: \"b7a1e678-e517-4b2b-9915-323590b80ff5\") " Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.660003 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7a1e678-e517-4b2b-9915-323590b80ff5-dns-svc\") pod \"b7a1e678-e517-4b2b-9915-323590b80ff5\" (UID: \"b7a1e678-e517-4b2b-9915-323590b80ff5\") " Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.660220 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwl2g\" (UniqueName: \"kubernetes.io/projected/b7a1e678-e517-4b2b-9915-323590b80ff5-kube-api-access-zwl2g\") pod \"b7a1e678-e517-4b2b-9915-323590b80ff5\" (UID: \"b7a1e678-e517-4b2b-9915-323590b80ff5\") " Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.661409 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7a1e678-e517-4b2b-9915-323590b80ff5-config" (OuterVolumeSpecName: "config") pod "b7a1e678-e517-4b2b-9915-323590b80ff5" (UID: "b7a1e678-e517-4b2b-9915-323590b80ff5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.661730 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7a1e678-e517-4b2b-9915-323590b80ff5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b7a1e678-e517-4b2b-9915-323590b80ff5" (UID: "b7a1e678-e517-4b2b-9915-323590b80ff5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.679545 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7a1e678-e517-4b2b-9915-323590b80ff5-kube-api-access-zwl2g" (OuterVolumeSpecName: "kube-api-access-zwl2g") pod "b7a1e678-e517-4b2b-9915-323590b80ff5" (UID: "b7a1e678-e517-4b2b-9915-323590b80ff5"). InnerVolumeSpecName "kube-api-access-zwl2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.767086 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwl2g\" (UniqueName: \"kubernetes.io/projected/b7a1e678-e517-4b2b-9915-323590b80ff5-kube-api-access-zwl2g\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.767444 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a1e678-e517-4b2b-9915-323590b80ff5-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:11 crc kubenswrapper[4629]: I1211 08:56:11.767457 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7a1e678-e517-4b2b-9915-323590b80ff5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.000432 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-pf5fp"] Dec 11 08:56:12 crc kubenswrapper[4629]: W1211 08:56:12.045691 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod902107e5_7c95_4d45_b298_93b32294cfb6.slice/crio-6f302e230d50ecb7a5b4ebca8c23416b1d1d1f778f554ceb0b744be4339a2079 WatchSource:0}: Error finding container 6f302e230d50ecb7a5b4ebca8c23416b1d1d1f778f554ceb0b744be4339a2079: Status 404 returned error can't find the container with id 6f302e230d50ecb7a5b4ebca8c23416b1d1d1f778f554ceb0b744be4339a2079 Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.128055 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vzhn9"] Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.227288 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" event={"ID":"902107e5-7c95-4d45-b298-93b32294cfb6","Type":"ContainerStarted","Data":"6f302e230d50ecb7a5b4ebca8c23416b1d1d1f778f554ceb0b744be4339a2079"} Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.227333 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d2811a7c-5ac9-4db5-81d3-61771795a87e","Type":"ContainerStarted","Data":"c267e69327de510172bec0ce8ef2ded8f17011c1c607739aa76ec7ffa8a6915b"} Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.227347 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"762352f0-3934-4d44-b846-fa922ffedd1d","Type":"ContainerStarted","Data":"57e4b3d2780a7f0bd1c2c024806eda3f95374863378664f7082ff008eaa8c030"} Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.232314 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vzhn9" event={"ID":"8f807dbd-baed-4ee5-b422-8a3a4e2d045f","Type":"ContainerStarted","Data":"466574b805fe4d265b857a8a9d8b9e868f545af40c94051c14c08bee238c0f06"} Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.251476 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"0ccfb64c-586b-4381-96b7-5846746d6433","Type":"ContainerStarted","Data":"5d8f8833ab87a5103ddb18d318f5b1f3cc37d56d670f53d39f5d54e0eee12767"} Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.253940 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v9ctv" event={"ID":"5a52781a-0193-4e3a-b0d3-4ee50d507f1a","Type":"ContainerStarted","Data":"6ee35638c4abb5de265ec24ef033f2d6b3ad0cf38546f9a4d8ffce124b4333bb"} Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.262347 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ab910fe2-4e06-4174-9638-e9b920c5bff9","Type":"ContainerStarted","Data":"f668a273774864322927decaf4a867c109f127fe81bc8ffd556e2fc949160f01"} Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.282925 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dxs2v" Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.283098 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.283124 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-hpgr4" event={"ID":"b7a1e678-e517-4b2b-9915-323590b80ff5","Type":"ContainerDied","Data":"7ff762e75120b3985c26704b3698ac8492da02a4e3ce5ce5d3f609893a56d572"} Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.316060 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-vdhgz"] Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.417539 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dxs2v"] Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.431271 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dxs2v"] Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.458166 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-hpgr4"] Dec 11 08:56:12 crc kubenswrapper[4629]: I1211 08:56:12.477468 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-hpgr4"] Dec 11 08:56:13 crc kubenswrapper[4629]: I1211 08:56:13.328112 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8lfjz" event={"ID":"708ce743-d291-41e4-8c15-a9ad1cda240e","Type":"ContainerStarted","Data":"dd149bd75d2d412c10e4abf8d933b47ebe9c2e74bddc6be6d056e48046e0ac49"} Dec 11 08:56:13 crc kubenswrapper[4629]: I1211 08:56:13.328504 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-8lfjz" Dec 11 08:56:13 crc kubenswrapper[4629]: I1211 08:56:13.334893 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-vdhgz" event={"ID":"6ddc853e-11fc-4d38-b550-8b42d1997c2b","Type":"ContainerStarted","Data":"b1a220e949c42352a487a5a40b2c71c4c720628b8c01114fc8cf91a5e22442ad"} Dec 11 08:56:13 crc kubenswrapper[4629]: I1211 08:56:13.338758 4629 generic.go:334] "Generic (PLEG): container finished" podID="5a52781a-0193-4e3a-b0d3-4ee50d507f1a" containerID="6ee35638c4abb5de265ec24ef033f2d6b3ad0cf38546f9a4d8ffce124b4333bb" exitCode=0 Dec 11 08:56:13 crc kubenswrapper[4629]: I1211 08:56:13.339569 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v9ctv" event={"ID":"5a52781a-0193-4e3a-b0d3-4ee50d507f1a","Type":"ContainerDied","Data":"6ee35638c4abb5de265ec24ef033f2d6b3ad0cf38546f9a4d8ffce124b4333bb"} Dec 11 08:56:13 crc kubenswrapper[4629]: I1211 08:56:13.356939 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-8lfjz" podStartSLOduration=17.336591709 podStartE2EDuration="27.356921837s" podCreationTimestamp="2025-12-11 08:55:46 +0000 UTC" firstStartedPulling="2025-12-11 08:56:00.523361567 +0000 UTC m=+1148.588777175" lastFinishedPulling="2025-12-11 08:56:10.543691695 +0000 UTC m=+1158.609107303" observedRunningTime="2025-12-11 08:56:13.353412226 +0000 UTC m=+1161.418827834" watchObservedRunningTime="2025-12-11 08:56:13.356921837 +0000 UTC m=+1161.422337445" Dec 11 08:56:14 crc kubenswrapper[4629]: I1211 08:56:14.213433 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b501db9-e5f4-484c-beb6-0085345af291" path="/var/lib/kubelet/pods/6b501db9-e5f4-484c-beb6-0085345af291/volumes" Dec 11 08:56:14 crc kubenswrapper[4629]: I1211 08:56:14.214381 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7a1e678-e517-4b2b-9915-323590b80ff5" path="/var/lib/kubelet/pods/b7a1e678-e517-4b2b-9915-323590b80ff5/volumes" Dec 11 08:56:14 crc kubenswrapper[4629]: I1211 08:56:14.365087 4629 generic.go:334] "Generic (PLEG): container finished" podID="902107e5-7c95-4d45-b298-93b32294cfb6" containerID="f06d18433520e72f0f8a493e1cf9070a124fc80338a721867cfca2458b50b652" exitCode=0 Dec 11 08:56:14 crc kubenswrapper[4629]: I1211 08:56:14.365152 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" event={"ID":"902107e5-7c95-4d45-b298-93b32294cfb6","Type":"ContainerDied","Data":"f06d18433520e72f0f8a493e1cf9070a124fc80338a721867cfca2458b50b652"} Dec 11 08:56:14 crc kubenswrapper[4629]: I1211 08:56:14.396890 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2594a86c-caf2-4874-a415-4e6fda88db08","Type":"ContainerStarted","Data":"1cd05b725be1d7583bf973ed8dba0ef5f55f3f153bef4abf0bc35e8be720622c"} Dec 11 08:56:14 crc kubenswrapper[4629]: I1211 08:56:14.398156 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 11 08:56:14 crc kubenswrapper[4629]: I1211 08:56:14.413412 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v9ctv" event={"ID":"5a52781a-0193-4e3a-b0d3-4ee50d507f1a","Type":"ContainerStarted","Data":"efe949339b7d07bbfaf332b6d784a0760a36b5bf03b6be96290f38febea7dd21"} Dec 11 08:56:14 crc kubenswrapper[4629]: I1211 08:56:14.413458 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v9ctv" event={"ID":"5a52781a-0193-4e3a-b0d3-4ee50d507f1a","Type":"ContainerStarted","Data":"de8850870cd37b1e24af2cb1240f226f88de1efef2d95c14e67b6fe092d16648"} Dec 11 08:56:14 crc kubenswrapper[4629]: I1211 08:56:14.413478 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:56:14 crc kubenswrapper[4629]: I1211 08:56:14.413510 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:56:14 crc kubenswrapper[4629]: I1211 08:56:14.451381 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=19.972335621 podStartE2EDuration="32.45135732s" podCreationTimestamp="2025-12-11 08:55:42 +0000 UTC" firstStartedPulling="2025-12-11 08:56:00.742654372 +0000 UTC m=+1148.808069980" lastFinishedPulling="2025-12-11 08:56:13.221676071 +0000 UTC m=+1161.287091679" observedRunningTime="2025-12-11 08:56:14.422516425 +0000 UTC m=+1162.487932043" watchObservedRunningTime="2025-12-11 08:56:14.45135732 +0000 UTC m=+1162.516772928" Dec 11 08:56:14 crc kubenswrapper[4629]: I1211 08:56:14.482678 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-v9ctv" podStartSLOduration=17.998552291 podStartE2EDuration="27.482654553s" podCreationTimestamp="2025-12-11 08:55:47 +0000 UTC" firstStartedPulling="2025-12-11 08:56:01.041524016 +0000 UTC m=+1149.106939624" lastFinishedPulling="2025-12-11 08:56:10.525626278 +0000 UTC m=+1158.591041886" observedRunningTime="2025-12-11 08:56:14.450623188 +0000 UTC m=+1162.516038796" watchObservedRunningTime="2025-12-11 08:56:14.482654553 +0000 UTC m=+1162.548070171" Dec 11 08:56:15 crc kubenswrapper[4629]: I1211 08:56:15.423323 4629 generic.go:334] "Generic (PLEG): container finished" podID="6ddc853e-11fc-4d38-b550-8b42d1997c2b" containerID="b3c53854322922e062a73b447c888cb0d882bd1f611f87a1ccc01784ba16cca4" exitCode=0 Dec 11 08:56:15 crc kubenswrapper[4629]: I1211 08:56:15.423439 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-vdhgz" event={"ID":"6ddc853e-11fc-4d38-b550-8b42d1997c2b","Type":"ContainerDied","Data":"b3c53854322922e062a73b447c888cb0d882bd1f611f87a1ccc01784ba16cca4"} Dec 11 08:56:15 crc kubenswrapper[4629]: I1211 08:56:15.440705 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" event={"ID":"902107e5-7c95-4d45-b298-93b32294cfb6","Type":"ContainerStarted","Data":"1b59c5085464510b99ea1e50831fe1c8ae61536de792555534305bdbb38d130f"} Dec 11 08:56:15 crc kubenswrapper[4629]: I1211 08:56:15.442114 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:16 crc kubenswrapper[4629]: I1211 08:56:16.488443 4629 generic.go:334] "Generic (PLEG): container finished" podID="ab910fe2-4e06-4174-9638-e9b920c5bff9" containerID="f668a273774864322927decaf4a867c109f127fe81bc8ffd556e2fc949160f01" exitCode=0 Dec 11 08:56:16 crc kubenswrapper[4629]: I1211 08:56:16.488689 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ab910fe2-4e06-4174-9638-e9b920c5bff9","Type":"ContainerDied","Data":"f668a273774864322927decaf4a867c109f127fe81bc8ffd556e2fc949160f01"} Dec 11 08:56:16 crc kubenswrapper[4629]: I1211 08:56:16.493165 4629 generic.go:334] "Generic (PLEG): container finished" podID="762352f0-3934-4d44-b846-fa922ffedd1d" containerID="57e4b3d2780a7f0bd1c2c024806eda3f95374863378664f7082ff008eaa8c030" exitCode=0 Dec 11 08:56:16 crc kubenswrapper[4629]: I1211 08:56:16.495172 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"762352f0-3934-4d44-b846-fa922ffedd1d","Type":"ContainerDied","Data":"57e4b3d2780a7f0bd1c2c024806eda3f95374863378664f7082ff008eaa8c030"} Dec 11 08:56:16 crc kubenswrapper[4629]: I1211 08:56:16.532746 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" podStartSLOduration=5.247159409 podStartE2EDuration="6.532538237s" podCreationTimestamp="2025-12-11 08:56:10 +0000 UTC" firstStartedPulling="2025-12-11 08:56:12.059686386 +0000 UTC m=+1160.125101994" lastFinishedPulling="2025-12-11 08:56:13.345065214 +0000 UTC m=+1161.410480822" observedRunningTime="2025-12-11 08:56:15.465463612 +0000 UTC m=+1163.530879230" watchObservedRunningTime="2025-12-11 08:56:16.532538237 +0000 UTC m=+1164.597953855" Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.517685 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ab910fe2-4e06-4174-9638-e9b920c5bff9","Type":"ContainerStarted","Data":"3ed2bac08cf37091a5a2fe5058743d0bdac1d6ca35bf544528ea8091d91de4ae"} Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.519486 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d2811a7c-5ac9-4db5-81d3-61771795a87e","Type":"ContainerStarted","Data":"225981d85a8a35aa8bcae62ed06a52839f27fc386367a84359fe2a73609a31e1"} Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.521171 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"762352f0-3934-4d44-b846-fa922ffedd1d","Type":"ContainerStarted","Data":"62bf60d36fbc6ce97fafe4ad89bee09bdce1c329a9aa703008b475a9545b20f6"} Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.522479 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vzhn9" event={"ID":"8f807dbd-baed-4ee5-b422-8a3a4e2d045f","Type":"ContainerStarted","Data":"05a8eaac9c3f068d0a9eda23e5b64d6539cd7a424aa6a84afecbf3e5f885470e"} Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.524506 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-vdhgz" event={"ID":"6ddc853e-11fc-4d38-b550-8b42d1997c2b","Type":"ContainerStarted","Data":"7cd6e7708100c30d73d22b66cf3e0b6d2a282c2a01cefc223709bf698b55984c"} Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.524649 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.526187 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"0ccfb64c-586b-4381-96b7-5846746d6433","Type":"ContainerStarted","Data":"ad4241faccf48515516ff851e082188d0da41917f6dde558073f9452c71d95ac"} Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.561215 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=18.986391445 podStartE2EDuration="34.561188015s" podCreationTimestamp="2025-12-11 08:55:44 +0000 UTC" firstStartedPulling="2025-12-11 08:56:02.312129711 +0000 UTC m=+1150.377545319" lastFinishedPulling="2025-12-11 08:56:17.886926281 +0000 UTC m=+1165.952341889" observedRunningTime="2025-12-11 08:56:18.557088657 +0000 UTC m=+1166.622504295" watchObservedRunningTime="2025-12-11 08:56:18.561188015 +0000 UTC m=+1166.626603623" Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.565502 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=30.830518623 podStartE2EDuration="40.56548137s" podCreationTimestamp="2025-12-11 08:55:38 +0000 UTC" firstStartedPulling="2025-12-11 08:56:00.491327131 +0000 UTC m=+1148.556742739" lastFinishedPulling="2025-12-11 08:56:10.226289878 +0000 UTC m=+1158.291705486" observedRunningTime="2025-12-11 08:56:18.541107694 +0000 UTC m=+1166.606523312" watchObservedRunningTime="2025-12-11 08:56:18.56548137 +0000 UTC m=+1166.630896978" Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.581656 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-vzhn9" podStartSLOduration=2.860605545 podStartE2EDuration="8.581634117s" podCreationTimestamp="2025-12-11 08:56:10 +0000 UTC" firstStartedPulling="2025-12-11 08:56:12.147018317 +0000 UTC m=+1160.212433925" lastFinishedPulling="2025-12-11 08:56:17.868046889 +0000 UTC m=+1165.933462497" observedRunningTime="2025-12-11 08:56:18.573787551 +0000 UTC m=+1166.639203159" watchObservedRunningTime="2025-12-11 08:56:18.581634117 +0000 UTC m=+1166.647049725" Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.602610 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-vdhgz" podStartSLOduration=6.150223823 podStartE2EDuration="8.602594404s" podCreationTimestamp="2025-12-11 08:56:10 +0000 UTC" firstStartedPulling="2025-12-11 08:56:12.316866331 +0000 UTC m=+1160.382281939" lastFinishedPulling="2025-12-11 08:56:14.769236912 +0000 UTC m=+1162.834652520" observedRunningTime="2025-12-11 08:56:18.595804962 +0000 UTC m=+1166.661220570" watchObservedRunningTime="2025-12-11 08:56:18.602594404 +0000 UTC m=+1166.668010012" Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.656610 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=13.875059552 podStartE2EDuration="30.656587269s" podCreationTimestamp="2025-12-11 08:55:48 +0000 UTC" firstStartedPulling="2025-12-11 08:56:01.239106971 +0000 UTC m=+1149.304522579" lastFinishedPulling="2025-12-11 08:56:18.020634688 +0000 UTC m=+1166.086050296" observedRunningTime="2025-12-11 08:56:18.654489644 +0000 UTC m=+1166.719905262" watchObservedRunningTime="2025-12-11 08:56:18.656587269 +0000 UTC m=+1166.722002877" Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.660890 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=31.636626162 podStartE2EDuration="41.660871314s" podCreationTimestamp="2025-12-11 08:55:37 +0000 UTC" firstStartedPulling="2025-12-11 08:56:00.990566996 +0000 UTC m=+1149.055982604" lastFinishedPulling="2025-12-11 08:56:11.014812148 +0000 UTC m=+1159.080227756" observedRunningTime="2025-12-11 08:56:18.631358937 +0000 UTC m=+1166.696774545" watchObservedRunningTime="2025-12-11 08:56:18.660871314 +0000 UTC m=+1166.726286922" Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.872500 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 11 08:56:18 crc kubenswrapper[4629]: I1211 08:56:18.872898 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 11 08:56:19 crc kubenswrapper[4629]: I1211 08:56:19.106506 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 11 08:56:19 crc kubenswrapper[4629]: I1211 08:56:19.142784 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 11 08:56:19 crc kubenswrapper[4629]: I1211 08:56:19.533470 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 11 08:56:19 crc kubenswrapper[4629]: I1211 08:56:19.576073 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.086189 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.086238 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.412221 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.412285 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.458293 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.485730 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.486038 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.579216 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.792664 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.794119 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.796817 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.796960 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.798024 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-ngbm6" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.809130 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.817036 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.901550 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8967f98-185a-4b7e-a43b-ff0555430bc1-scripts\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.901907 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f8967f98-185a-4b7e-a43b-ff0555430bc1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.902058 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qx2x\" (UniqueName: \"kubernetes.io/projected/f8967f98-185a-4b7e-a43b-ff0555430bc1-kube-api-access-6qx2x\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.902203 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8967f98-185a-4b7e-a43b-ff0555430bc1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.902318 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8967f98-185a-4b7e-a43b-ff0555430bc1-config\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.902452 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8967f98-185a-4b7e-a43b-ff0555430bc1-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:20 crc kubenswrapper[4629]: I1211 08:56:20.902567 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8967f98-185a-4b7e-a43b-ff0555430bc1-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.005250 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f8967f98-185a-4b7e-a43b-ff0555430bc1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.005333 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qx2x\" (UniqueName: \"kubernetes.io/projected/f8967f98-185a-4b7e-a43b-ff0555430bc1-kube-api-access-6qx2x\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.005401 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8967f98-185a-4b7e-a43b-ff0555430bc1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.005433 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8967f98-185a-4b7e-a43b-ff0555430bc1-config\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.005468 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8967f98-185a-4b7e-a43b-ff0555430bc1-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.005511 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8967f98-185a-4b7e-a43b-ff0555430bc1-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.005572 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8967f98-185a-4b7e-a43b-ff0555430bc1-scripts\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.006610 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f8967f98-185a-4b7e-a43b-ff0555430bc1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.006806 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8967f98-185a-4b7e-a43b-ff0555430bc1-scripts\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.006919 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8967f98-185a-4b7e-a43b-ff0555430bc1-config\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.015360 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8967f98-185a-4b7e-a43b-ff0555430bc1-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.027985 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.029563 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8967f98-185a-4b7e-a43b-ff0555430bc1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.031433 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qx2x\" (UniqueName: \"kubernetes.io/projected/f8967f98-185a-4b7e-a43b-ff0555430bc1-kube-api-access-6qx2x\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.036118 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8967f98-185a-4b7e-a43b-ff0555430bc1-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f8967f98-185a-4b7e-a43b-ff0555430bc1\") " pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.121022 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 11 08:56:21 crc kubenswrapper[4629]: I1211 08:56:21.674720 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 11 08:56:22 crc kubenswrapper[4629]: I1211 08:56:22.566636 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f8967f98-185a-4b7e-a43b-ff0555430bc1","Type":"ContainerStarted","Data":"4efa726a3647ede237c3ed44d3082961134af10ff37f42471f3c4fdc473e90bb"} Dec 11 08:56:23 crc kubenswrapper[4629]: I1211 08:56:23.015057 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 11 08:56:25 crc kubenswrapper[4629]: I1211 08:56:25.588073 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f8967f98-185a-4b7e-a43b-ff0555430bc1","Type":"ContainerStarted","Data":"1083c3cd8b4b21e983db79785a40810facf265417002018126b8a4c828080d4a"} Dec 11 08:56:26 crc kubenswrapper[4629]: I1211 08:56:26.255070 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:56:26 crc kubenswrapper[4629]: I1211 08:56:26.332422 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-pf5fp"] Dec 11 08:56:26 crc kubenswrapper[4629]: I1211 08:56:26.332652 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" podUID="902107e5-7c95-4d45-b298-93b32294cfb6" containerName="dnsmasq-dns" containerID="cri-o://1b59c5085464510b99ea1e50831fe1c8ae61536de792555534305bdbb38d130f" gracePeriod=10 Dec 11 08:56:26 crc kubenswrapper[4629]: I1211 08:56:26.600472 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f8967f98-185a-4b7e-a43b-ff0555430bc1","Type":"ContainerStarted","Data":"11c4b3b29a3f0f001682da57ff4513a9131e4faf874330d12439cd283760b14a"} Dec 11 08:56:26 crc kubenswrapper[4629]: I1211 08:56:26.602095 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 11 08:56:26 crc kubenswrapper[4629]: I1211 08:56:26.606384 4629 generic.go:334] "Generic (PLEG): container finished" podID="902107e5-7c95-4d45-b298-93b32294cfb6" containerID="1b59c5085464510b99ea1e50831fe1c8ae61536de792555534305bdbb38d130f" exitCode=0 Dec 11 08:56:26 crc kubenswrapper[4629]: I1211 08:56:26.606435 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" event={"ID":"902107e5-7c95-4d45-b298-93b32294cfb6","Type":"ContainerDied","Data":"1b59c5085464510b99ea1e50831fe1c8ae61536de792555534305bdbb38d130f"} Dec 11 08:56:26 crc kubenswrapper[4629]: I1211 08:56:26.634769 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.051705792 podStartE2EDuration="6.634750145s" podCreationTimestamp="2025-12-11 08:56:20 +0000 UTC" firstStartedPulling="2025-12-11 08:56:21.678415078 +0000 UTC m=+1169.743830686" lastFinishedPulling="2025-12-11 08:56:25.261459431 +0000 UTC m=+1173.326875039" observedRunningTime="2025-12-11 08:56:26.630615496 +0000 UTC m=+1174.696031104" watchObservedRunningTime="2025-12-11 08:56:26.634750145 +0000 UTC m=+1174.700165753" Dec 11 08:56:26 crc kubenswrapper[4629]: I1211 08:56:26.894670 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:26 crc kubenswrapper[4629]: I1211 08:56:26.953807 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.008876 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.043248 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-ovsdbserver-sb\") pod \"902107e5-7c95-4d45-b298-93b32294cfb6\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.043296 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-config\") pod \"902107e5-7c95-4d45-b298-93b32294cfb6\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.043360 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-dns-svc\") pod \"902107e5-7c95-4d45-b298-93b32294cfb6\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.043441 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9fk9\" (UniqueName: \"kubernetes.io/projected/902107e5-7c95-4d45-b298-93b32294cfb6-kube-api-access-h9fk9\") pod \"902107e5-7c95-4d45-b298-93b32294cfb6\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.056260 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.058086 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/902107e5-7c95-4d45-b298-93b32294cfb6-kube-api-access-h9fk9" (OuterVolumeSpecName: "kube-api-access-h9fk9") pod "902107e5-7c95-4d45-b298-93b32294cfb6" (UID: "902107e5-7c95-4d45-b298-93b32294cfb6"). InnerVolumeSpecName "kube-api-access-h9fk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.102009 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "902107e5-7c95-4d45-b298-93b32294cfb6" (UID: "902107e5-7c95-4d45-b298-93b32294cfb6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:27 crc kubenswrapper[4629]: E1211 08:56:27.118690 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-config podName:902107e5-7c95-4d45-b298-93b32294cfb6 nodeName:}" failed. No retries permitted until 2025-12-11 08:56:27.618434698 +0000 UTC m=+1175.683850306 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-config") pod "902107e5-7c95-4d45-b298-93b32294cfb6" (UID: "902107e5-7c95-4d45-b298-93b32294cfb6") : error deleting /var/lib/kubelet/pods/902107e5-7c95-4d45-b298-93b32294cfb6/volume-subpaths: remove /var/lib/kubelet/pods/902107e5-7c95-4d45-b298-93b32294cfb6/volume-subpaths: no such file or directory Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.119298 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "902107e5-7c95-4d45-b298-93b32294cfb6" (UID: "902107e5-7c95-4d45-b298-93b32294cfb6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.145392 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.145427 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.145436 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9fk9\" (UniqueName: \"kubernetes.io/projected/902107e5-7c95-4d45-b298-93b32294cfb6-kube-api-access-h9fk9\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.153865 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.615669 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" event={"ID":"902107e5-7c95-4d45-b298-93b32294cfb6","Type":"ContainerDied","Data":"6f302e230d50ecb7a5b4ebca8c23416b1d1d1f778f554ceb0b744be4339a2079"} Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.616149 4629 scope.go:117] "RemoveContainer" containerID="1b59c5085464510b99ea1e50831fe1c8ae61536de792555534305bdbb38d130f" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.615749 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-pf5fp" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.637167 4629 scope.go:117] "RemoveContainer" containerID="f06d18433520e72f0f8a493e1cf9070a124fc80338a721867cfca2458b50b652" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.653539 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-config\") pod \"902107e5-7c95-4d45-b298-93b32294cfb6\" (UID: \"902107e5-7c95-4d45-b298-93b32294cfb6\") " Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.654039 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-config" (OuterVolumeSpecName: "config") pod "902107e5-7c95-4d45-b298-93b32294cfb6" (UID: "902107e5-7c95-4d45-b298-93b32294cfb6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.654256 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/902107e5-7c95-4d45-b298-93b32294cfb6-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.955752 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-pf5fp"] Dec 11 08:56:27 crc kubenswrapper[4629]: I1211 08:56:27.966979 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-pf5fp"] Dec 11 08:56:28 crc kubenswrapper[4629]: E1211 08:56:28.083680 4629 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod902107e5_7c95_4d45_b298_93b32294cfb6.slice\": RecentStats: unable to find data in memory cache]" Dec 11 08:56:28 crc kubenswrapper[4629]: I1211 08:56:28.207247 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="902107e5-7c95-4d45-b298-93b32294cfb6" path="/var/lib/kubelet/pods/902107e5-7c95-4d45-b298-93b32294cfb6/volumes" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.471260 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-tzqff"] Dec 11 08:56:30 crc kubenswrapper[4629]: E1211 08:56:30.471947 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="902107e5-7c95-4d45-b298-93b32294cfb6" containerName="init" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.471964 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="902107e5-7c95-4d45-b298-93b32294cfb6" containerName="init" Dec 11 08:56:30 crc kubenswrapper[4629]: E1211 08:56:30.471999 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="902107e5-7c95-4d45-b298-93b32294cfb6" containerName="dnsmasq-dns" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.472007 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="902107e5-7c95-4d45-b298-93b32294cfb6" containerName="dnsmasq-dns" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.472196 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="902107e5-7c95-4d45-b298-93b32294cfb6" containerName="dnsmasq-dns" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.472837 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tzqff" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.484102 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-tzqff"] Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.575422 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-005f-account-create-update-d7fhm"] Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.576450 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-005f-account-create-update-d7fhm" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.579169 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.596887 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-005f-account-create-update-d7fhm"] Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.610115 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdmpr\" (UniqueName: \"kubernetes.io/projected/54cce582-17c9-40d7-92f1-8a7c05a8ca57-kube-api-access-mdmpr\") pod \"keystone-db-create-tzqff\" (UID: \"54cce582-17c9-40d7-92f1-8a7c05a8ca57\") " pod="openstack/keystone-db-create-tzqff" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.610165 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54cce582-17c9-40d7-92f1-8a7c05a8ca57-operator-scripts\") pod \"keystone-db-create-tzqff\" (UID: \"54cce582-17c9-40d7-92f1-8a7c05a8ca57\") " pod="openstack/keystone-db-create-tzqff" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.711940 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdmpr\" (UniqueName: \"kubernetes.io/projected/54cce582-17c9-40d7-92f1-8a7c05a8ca57-kube-api-access-mdmpr\") pod \"keystone-db-create-tzqff\" (UID: \"54cce582-17c9-40d7-92f1-8a7c05a8ca57\") " pod="openstack/keystone-db-create-tzqff" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.711988 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54cce582-17c9-40d7-92f1-8a7c05a8ca57-operator-scripts\") pod \"keystone-db-create-tzqff\" (UID: \"54cce582-17c9-40d7-92f1-8a7c05a8ca57\") " pod="openstack/keystone-db-create-tzqff" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.712023 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48b7b556-b31c-4a94-aa25-3a26ba8d07dd-operator-scripts\") pod \"keystone-005f-account-create-update-d7fhm\" (UID: \"48b7b556-b31c-4a94-aa25-3a26ba8d07dd\") " pod="openstack/keystone-005f-account-create-update-d7fhm" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.712055 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwm6b\" (UniqueName: \"kubernetes.io/projected/48b7b556-b31c-4a94-aa25-3a26ba8d07dd-kube-api-access-fwm6b\") pod \"keystone-005f-account-create-update-d7fhm\" (UID: \"48b7b556-b31c-4a94-aa25-3a26ba8d07dd\") " pod="openstack/keystone-005f-account-create-update-d7fhm" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.712869 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54cce582-17c9-40d7-92f1-8a7c05a8ca57-operator-scripts\") pod \"keystone-db-create-tzqff\" (UID: \"54cce582-17c9-40d7-92f1-8a7c05a8ca57\") " pod="openstack/keystone-db-create-tzqff" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.740067 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdmpr\" (UniqueName: \"kubernetes.io/projected/54cce582-17c9-40d7-92f1-8a7c05a8ca57-kube-api-access-mdmpr\") pod \"keystone-db-create-tzqff\" (UID: \"54cce582-17c9-40d7-92f1-8a7c05a8ca57\") " pod="openstack/keystone-db-create-tzqff" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.777144 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-2sknt"] Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.778419 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2sknt" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.786441 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2sknt"] Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.791687 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tzqff" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.814191 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48b7b556-b31c-4a94-aa25-3a26ba8d07dd-operator-scripts\") pod \"keystone-005f-account-create-update-d7fhm\" (UID: \"48b7b556-b31c-4a94-aa25-3a26ba8d07dd\") " pod="openstack/keystone-005f-account-create-update-d7fhm" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.814261 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwm6b\" (UniqueName: \"kubernetes.io/projected/48b7b556-b31c-4a94-aa25-3a26ba8d07dd-kube-api-access-fwm6b\") pod \"keystone-005f-account-create-update-d7fhm\" (UID: \"48b7b556-b31c-4a94-aa25-3a26ba8d07dd\") " pod="openstack/keystone-005f-account-create-update-d7fhm" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.815195 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48b7b556-b31c-4a94-aa25-3a26ba8d07dd-operator-scripts\") pod \"keystone-005f-account-create-update-d7fhm\" (UID: \"48b7b556-b31c-4a94-aa25-3a26ba8d07dd\") " pod="openstack/keystone-005f-account-create-update-d7fhm" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.850812 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwm6b\" (UniqueName: \"kubernetes.io/projected/48b7b556-b31c-4a94-aa25-3a26ba8d07dd-kube-api-access-fwm6b\") pod \"keystone-005f-account-create-update-d7fhm\" (UID: \"48b7b556-b31c-4a94-aa25-3a26ba8d07dd\") " pod="openstack/keystone-005f-account-create-update-d7fhm" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.893337 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-005f-account-create-update-d7fhm" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.916240 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cnp2\" (UniqueName: \"kubernetes.io/projected/3da40552-275b-4c71-a2d5-2257bdad664d-kube-api-access-7cnp2\") pod \"placement-db-create-2sknt\" (UID: \"3da40552-275b-4c71-a2d5-2257bdad664d\") " pod="openstack/placement-db-create-2sknt" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.916290 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da40552-275b-4c71-a2d5-2257bdad664d-operator-scripts\") pod \"placement-db-create-2sknt\" (UID: \"3da40552-275b-4c71-a2d5-2257bdad664d\") " pod="openstack/placement-db-create-2sknt" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.918626 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a1e4-account-create-update-rlcdj"] Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.919623 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a1e4-account-create-update-rlcdj" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.925504 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 11 08:56:30 crc kubenswrapper[4629]: I1211 08:56:30.936292 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a1e4-account-create-update-rlcdj"] Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.019333 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbwkg\" (UniqueName: \"kubernetes.io/projected/0e81ddaf-9300-4a5b-b3cf-5f90928921a2-kube-api-access-qbwkg\") pod \"placement-a1e4-account-create-update-rlcdj\" (UID: \"0e81ddaf-9300-4a5b-b3cf-5f90928921a2\") " pod="openstack/placement-a1e4-account-create-update-rlcdj" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.019454 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cnp2\" (UniqueName: \"kubernetes.io/projected/3da40552-275b-4c71-a2d5-2257bdad664d-kube-api-access-7cnp2\") pod \"placement-db-create-2sknt\" (UID: \"3da40552-275b-4c71-a2d5-2257bdad664d\") " pod="openstack/placement-db-create-2sknt" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.019495 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da40552-275b-4c71-a2d5-2257bdad664d-operator-scripts\") pod \"placement-db-create-2sknt\" (UID: \"3da40552-275b-4c71-a2d5-2257bdad664d\") " pod="openstack/placement-db-create-2sknt" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.019576 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e81ddaf-9300-4a5b-b3cf-5f90928921a2-operator-scripts\") pod \"placement-a1e4-account-create-update-rlcdj\" (UID: \"0e81ddaf-9300-4a5b-b3cf-5f90928921a2\") " pod="openstack/placement-a1e4-account-create-update-rlcdj" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.033655 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da40552-275b-4c71-a2d5-2257bdad664d-operator-scripts\") pod \"placement-db-create-2sknt\" (UID: \"3da40552-275b-4c71-a2d5-2257bdad664d\") " pod="openstack/placement-db-create-2sknt" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.041012 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cnp2\" (UniqueName: \"kubernetes.io/projected/3da40552-275b-4c71-a2d5-2257bdad664d-kube-api-access-7cnp2\") pod \"placement-db-create-2sknt\" (UID: \"3da40552-275b-4c71-a2d5-2257bdad664d\") " pod="openstack/placement-db-create-2sknt" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.097289 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2sknt" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.124308 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e81ddaf-9300-4a5b-b3cf-5f90928921a2-operator-scripts\") pod \"placement-a1e4-account-create-update-rlcdj\" (UID: \"0e81ddaf-9300-4a5b-b3cf-5f90928921a2\") " pod="openstack/placement-a1e4-account-create-update-rlcdj" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.124591 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbwkg\" (UniqueName: \"kubernetes.io/projected/0e81ddaf-9300-4a5b-b3cf-5f90928921a2-kube-api-access-qbwkg\") pod \"placement-a1e4-account-create-update-rlcdj\" (UID: \"0e81ddaf-9300-4a5b-b3cf-5f90928921a2\") " pod="openstack/placement-a1e4-account-create-update-rlcdj" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.125165 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e81ddaf-9300-4a5b-b3cf-5f90928921a2-operator-scripts\") pod \"placement-a1e4-account-create-update-rlcdj\" (UID: \"0e81ddaf-9300-4a5b-b3cf-5f90928921a2\") " pod="openstack/placement-a1e4-account-create-update-rlcdj" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.157004 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-grhlk"] Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.160242 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-grhlk" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.163581 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbwkg\" (UniqueName: \"kubernetes.io/projected/0e81ddaf-9300-4a5b-b3cf-5f90928921a2-kube-api-access-qbwkg\") pod \"placement-a1e4-account-create-update-rlcdj\" (UID: \"0e81ddaf-9300-4a5b-b3cf-5f90928921a2\") " pod="openstack/placement-a1e4-account-create-update-rlcdj" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.200182 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-grhlk"] Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.266540 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-8495-account-create-update-5rsvg"] Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.268387 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8495-account-create-update-5rsvg" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.268925 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a1e4-account-create-update-rlcdj" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.271673 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.280297 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-8495-account-create-update-5rsvg"] Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.301916 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-tzqff"] Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.329587 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/623e0d38-cbc4-4261-ad37-319eb1704ce2-operator-scripts\") pod \"glance-db-create-grhlk\" (UID: \"623e0d38-cbc4-4261-ad37-319eb1704ce2\") " pod="openstack/glance-db-create-grhlk" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.329666 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drqzz\" (UniqueName: \"kubernetes.io/projected/623e0d38-cbc4-4261-ad37-319eb1704ce2-kube-api-access-drqzz\") pod \"glance-db-create-grhlk\" (UID: \"623e0d38-cbc4-4261-ad37-319eb1704ce2\") " pod="openstack/glance-db-create-grhlk" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.413745 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-005f-account-create-update-d7fhm"] Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.430785 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac-operator-scripts\") pod \"glance-8495-account-create-update-5rsvg\" (UID: \"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac\") " pod="openstack/glance-8495-account-create-update-5rsvg" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.430909 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btk2q\" (UniqueName: \"kubernetes.io/projected/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac-kube-api-access-btk2q\") pod \"glance-8495-account-create-update-5rsvg\" (UID: \"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac\") " pod="openstack/glance-8495-account-create-update-5rsvg" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.431016 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/623e0d38-cbc4-4261-ad37-319eb1704ce2-operator-scripts\") pod \"glance-db-create-grhlk\" (UID: \"623e0d38-cbc4-4261-ad37-319eb1704ce2\") " pod="openstack/glance-db-create-grhlk" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.431052 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drqzz\" (UniqueName: \"kubernetes.io/projected/623e0d38-cbc4-4261-ad37-319eb1704ce2-kube-api-access-drqzz\") pod \"glance-db-create-grhlk\" (UID: \"623e0d38-cbc4-4261-ad37-319eb1704ce2\") " pod="openstack/glance-db-create-grhlk" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.431922 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/623e0d38-cbc4-4261-ad37-319eb1704ce2-operator-scripts\") pod \"glance-db-create-grhlk\" (UID: \"623e0d38-cbc4-4261-ad37-319eb1704ce2\") " pod="openstack/glance-db-create-grhlk" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.463263 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drqzz\" (UniqueName: \"kubernetes.io/projected/623e0d38-cbc4-4261-ad37-319eb1704ce2-kube-api-access-drqzz\") pod \"glance-db-create-grhlk\" (UID: \"623e0d38-cbc4-4261-ad37-319eb1704ce2\") " pod="openstack/glance-db-create-grhlk" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.482634 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-grhlk" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.541537 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac-operator-scripts\") pod \"glance-8495-account-create-update-5rsvg\" (UID: \"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac\") " pod="openstack/glance-8495-account-create-update-5rsvg" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.541603 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btk2q\" (UniqueName: \"kubernetes.io/projected/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac-kube-api-access-btk2q\") pod \"glance-8495-account-create-update-5rsvg\" (UID: \"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac\") " pod="openstack/glance-8495-account-create-update-5rsvg" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.542443 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac-operator-scripts\") pod \"glance-8495-account-create-update-5rsvg\" (UID: \"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac\") " pod="openstack/glance-8495-account-create-update-5rsvg" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.575613 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btk2q\" (UniqueName: \"kubernetes.io/projected/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac-kube-api-access-btk2q\") pod \"glance-8495-account-create-update-5rsvg\" (UID: \"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac\") " pod="openstack/glance-8495-account-create-update-5rsvg" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.603708 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8495-account-create-update-5rsvg" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.653089 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2sknt"] Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.691225 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-005f-account-create-update-d7fhm" event={"ID":"48b7b556-b31c-4a94-aa25-3a26ba8d07dd","Type":"ContainerStarted","Data":"3c12541e46b5d0ab526e33aa06343d192ff389f9853811ffd416d091b4ac8a76"} Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.691343 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-005f-account-create-update-d7fhm" event={"ID":"48b7b556-b31c-4a94-aa25-3a26ba8d07dd","Type":"ContainerStarted","Data":"7cbe4f521a420f3411cb726f644db16f7f7ca3834b2411c4a030e058f41ab6f4"} Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.693628 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tzqff" event={"ID":"54cce582-17c9-40d7-92f1-8a7c05a8ca57","Type":"ContainerStarted","Data":"5f1bc94f5c7ffcf73fd858a15f2e7f06042ad0d3eceffee0499845e2b0a78215"} Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.693695 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tzqff" event={"ID":"54cce582-17c9-40d7-92f1-8a7c05a8ca57","Type":"ContainerStarted","Data":"9c52daac787d4422f6b69d8ad6a3c2d9831a9ca81550d36cb1d1eeac8f86dfee"} Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.715521 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-005f-account-create-update-d7fhm" podStartSLOduration=1.71550394 podStartE2EDuration="1.71550394s" podCreationTimestamp="2025-12-11 08:56:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:56:31.708742857 +0000 UTC m=+1179.774158465" watchObservedRunningTime="2025-12-11 08:56:31.71550394 +0000 UTC m=+1179.780919548" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.740638 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-tzqff" podStartSLOduration=1.7406175579999998 podStartE2EDuration="1.740617558s" podCreationTimestamp="2025-12-11 08:56:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:56:31.729272062 +0000 UTC m=+1179.794687670" watchObservedRunningTime="2025-12-11 08:56:31.740617558 +0000 UTC m=+1179.806033156" Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.786380 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a1e4-account-create-update-rlcdj"] Dec 11 08:56:31 crc kubenswrapper[4629]: W1211 08:56:31.843557 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e81ddaf_9300_4a5b_b3cf_5f90928921a2.slice/crio-8d53c854c685da347b58cece36dc768c56351977cc5f234772e0591ba5540a67 WatchSource:0}: Error finding container 8d53c854c685da347b58cece36dc768c56351977cc5f234772e0591ba5540a67: Status 404 returned error can't find the container with id 8d53c854c685da347b58cece36dc768c56351977cc5f234772e0591ba5540a67 Dec 11 08:56:31 crc kubenswrapper[4629]: I1211 08:56:31.933915 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-grhlk"] Dec 11 08:56:31 crc kubenswrapper[4629]: W1211 08:56:31.946924 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod623e0d38_cbc4_4261_ad37_319eb1704ce2.slice/crio-136b96aaf9d8f9e9022af40bc1c472e920980ad97288a1b568b2c7269c5a5ce6 WatchSource:0}: Error finding container 136b96aaf9d8f9e9022af40bc1c472e920980ad97288a1b568b2c7269c5a5ce6: Status 404 returned error can't find the container with id 136b96aaf9d8f9e9022af40bc1c472e920980ad97288a1b568b2c7269c5a5ce6 Dec 11 08:56:32 crc kubenswrapper[4629]: W1211 08:56:32.111886 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7f3b384_3bf7_46e1_9316_9c83bac9a8ac.slice/crio-79c5c82660811ea3cbbab07a33959038bdd5793bca634990137092bdb576147f WatchSource:0}: Error finding container 79c5c82660811ea3cbbab07a33959038bdd5793bca634990137092bdb576147f: Status 404 returned error can't find the container with id 79c5c82660811ea3cbbab07a33959038bdd5793bca634990137092bdb576147f Dec 11 08:56:32 crc kubenswrapper[4629]: I1211 08:56:32.112311 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-8495-account-create-update-5rsvg"] Dec 11 08:56:32 crc kubenswrapper[4629]: I1211 08:56:32.706152 4629 generic.go:334] "Generic (PLEG): container finished" podID="48b7b556-b31c-4a94-aa25-3a26ba8d07dd" containerID="3c12541e46b5d0ab526e33aa06343d192ff389f9853811ffd416d091b4ac8a76" exitCode=0 Dec 11 08:56:32 crc kubenswrapper[4629]: I1211 08:56:32.706213 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-005f-account-create-update-d7fhm" event={"ID":"48b7b556-b31c-4a94-aa25-3a26ba8d07dd","Type":"ContainerDied","Data":"3c12541e46b5d0ab526e33aa06343d192ff389f9853811ffd416d091b4ac8a76"} Dec 11 08:56:32 crc kubenswrapper[4629]: I1211 08:56:32.709498 4629 generic.go:334] "Generic (PLEG): container finished" podID="54cce582-17c9-40d7-92f1-8a7c05a8ca57" containerID="5f1bc94f5c7ffcf73fd858a15f2e7f06042ad0d3eceffee0499845e2b0a78215" exitCode=0 Dec 11 08:56:32 crc kubenswrapper[4629]: I1211 08:56:32.709568 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tzqff" event={"ID":"54cce582-17c9-40d7-92f1-8a7c05a8ca57","Type":"ContainerDied","Data":"5f1bc94f5c7ffcf73fd858a15f2e7f06042ad0d3eceffee0499845e2b0a78215"} Dec 11 08:56:32 crc kubenswrapper[4629]: I1211 08:56:32.712269 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-grhlk" event={"ID":"623e0d38-cbc4-4261-ad37-319eb1704ce2","Type":"ContainerStarted","Data":"136b96aaf9d8f9e9022af40bc1c472e920980ad97288a1b568b2c7269c5a5ce6"} Dec 11 08:56:32 crc kubenswrapper[4629]: I1211 08:56:32.713649 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2sknt" event={"ID":"3da40552-275b-4c71-a2d5-2257bdad664d","Type":"ContainerStarted","Data":"785bd98d77ffeb6600d497ac0ba9f99e63b9efe2fc58e603d37acd2ceb67126e"} Dec 11 08:56:32 crc kubenswrapper[4629]: I1211 08:56:32.715023 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8495-account-create-update-5rsvg" event={"ID":"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac","Type":"ContainerStarted","Data":"79c5c82660811ea3cbbab07a33959038bdd5793bca634990137092bdb576147f"} Dec 11 08:56:32 crc kubenswrapper[4629]: I1211 08:56:32.716135 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a1e4-account-create-update-rlcdj" event={"ID":"0e81ddaf-9300-4a5b-b3cf-5f90928921a2","Type":"ContainerStarted","Data":"8d53c854c685da347b58cece36dc768c56351977cc5f234772e0591ba5540a67"} Dec 11 08:56:33 crc kubenswrapper[4629]: I1211 08:56:33.723993 4629 generic.go:334] "Generic (PLEG): container finished" podID="623e0d38-cbc4-4261-ad37-319eb1704ce2" containerID="7fc7ce5deccc9a06d29ba025efbe3ef62969d3f1374f909f587815d813d69aa8" exitCode=0 Dec 11 08:56:33 crc kubenswrapper[4629]: I1211 08:56:33.724072 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-grhlk" event={"ID":"623e0d38-cbc4-4261-ad37-319eb1704ce2","Type":"ContainerDied","Data":"7fc7ce5deccc9a06d29ba025efbe3ef62969d3f1374f909f587815d813d69aa8"} Dec 11 08:56:33 crc kubenswrapper[4629]: I1211 08:56:33.725477 4629 generic.go:334] "Generic (PLEG): container finished" podID="0c633337-5e1d-438b-8b18-0acb62ec2da0" containerID="4a5a857a435fc978860280f4cbe148a7f4dd2ee9a3c7d2ab96099c938c4b272a" exitCode=0 Dec 11 08:56:33 crc kubenswrapper[4629]: I1211 08:56:33.725531 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0c633337-5e1d-438b-8b18-0acb62ec2da0","Type":"ContainerDied","Data":"4a5a857a435fc978860280f4cbe148a7f4dd2ee9a3c7d2ab96099c938c4b272a"} Dec 11 08:56:33 crc kubenswrapper[4629]: I1211 08:56:33.727407 4629 generic.go:334] "Generic (PLEG): container finished" podID="3da40552-275b-4c71-a2d5-2257bdad664d" containerID="2ed85de25378bb5b1589ce34a2970a8597348dce074d430820506e21f5e1f488" exitCode=0 Dec 11 08:56:33 crc kubenswrapper[4629]: I1211 08:56:33.727471 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2sknt" event={"ID":"3da40552-275b-4c71-a2d5-2257bdad664d","Type":"ContainerDied","Data":"2ed85de25378bb5b1589ce34a2970a8597348dce074d430820506e21f5e1f488"} Dec 11 08:56:33 crc kubenswrapper[4629]: I1211 08:56:33.733555 4629 generic.go:334] "Generic (PLEG): container finished" podID="b7f3b384-3bf7-46e1-9316-9c83bac9a8ac" containerID="0f16e1dcaba3fa2a7186d37a228a9d8fa0e4bddc8c6c74afa977e27d872a7fe7" exitCode=0 Dec 11 08:56:33 crc kubenswrapper[4629]: I1211 08:56:33.733641 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8495-account-create-update-5rsvg" event={"ID":"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac","Type":"ContainerDied","Data":"0f16e1dcaba3fa2a7186d37a228a9d8fa0e4bddc8c6c74afa977e27d872a7fe7"} Dec 11 08:56:33 crc kubenswrapper[4629]: I1211 08:56:33.742723 4629 generic.go:334] "Generic (PLEG): container finished" podID="0e81ddaf-9300-4a5b-b3cf-5f90928921a2" containerID="a107b196af4a4e5d373352c6c1f7fa8cc040a255a4a8eb20d4d7d4545d1e32a6" exitCode=0 Dec 11 08:56:33 crc kubenswrapper[4629]: I1211 08:56:33.742839 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a1e4-account-create-update-rlcdj" event={"ID":"0e81ddaf-9300-4a5b-b3cf-5f90928921a2","Type":"ContainerDied","Data":"a107b196af4a4e5d373352c6c1f7fa8cc040a255a4a8eb20d4d7d4545d1e32a6"} Dec 11 08:56:33 crc kubenswrapper[4629]: I1211 08:56:33.746989 4629 generic.go:334] "Generic (PLEG): container finished" podID="1a670dab-0f15-4572-b2a8-36ec2330adb2" containerID="d3ec5182b5c5afada44fd8ab808ed3ad84b1a4b2fbc5e3bd1a17034f8d925a9a" exitCode=0 Dec 11 08:56:33 crc kubenswrapper[4629]: I1211 08:56:33.747069 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1a670dab-0f15-4572-b2a8-36ec2330adb2","Type":"ContainerDied","Data":"d3ec5182b5c5afada44fd8ab808ed3ad84b1a4b2fbc5e3bd1a17034f8d925a9a"} Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.095000 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-005f-account-create-update-d7fhm" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.131330 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tzqff" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.191473 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48b7b556-b31c-4a94-aa25-3a26ba8d07dd-operator-scripts\") pod \"48b7b556-b31c-4a94-aa25-3a26ba8d07dd\" (UID: \"48b7b556-b31c-4a94-aa25-3a26ba8d07dd\") " Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.191796 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwm6b\" (UniqueName: \"kubernetes.io/projected/48b7b556-b31c-4a94-aa25-3a26ba8d07dd-kube-api-access-fwm6b\") pod \"48b7b556-b31c-4a94-aa25-3a26ba8d07dd\" (UID: \"48b7b556-b31c-4a94-aa25-3a26ba8d07dd\") " Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.192126 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48b7b556-b31c-4a94-aa25-3a26ba8d07dd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "48b7b556-b31c-4a94-aa25-3a26ba8d07dd" (UID: "48b7b556-b31c-4a94-aa25-3a26ba8d07dd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.192474 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48b7b556-b31c-4a94-aa25-3a26ba8d07dd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.196092 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48b7b556-b31c-4a94-aa25-3a26ba8d07dd-kube-api-access-fwm6b" (OuterVolumeSpecName: "kube-api-access-fwm6b") pod "48b7b556-b31c-4a94-aa25-3a26ba8d07dd" (UID: "48b7b556-b31c-4a94-aa25-3a26ba8d07dd"). InnerVolumeSpecName "kube-api-access-fwm6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.293147 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdmpr\" (UniqueName: \"kubernetes.io/projected/54cce582-17c9-40d7-92f1-8a7c05a8ca57-kube-api-access-mdmpr\") pod \"54cce582-17c9-40d7-92f1-8a7c05a8ca57\" (UID: \"54cce582-17c9-40d7-92f1-8a7c05a8ca57\") " Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.293374 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54cce582-17c9-40d7-92f1-8a7c05a8ca57-operator-scripts\") pod \"54cce582-17c9-40d7-92f1-8a7c05a8ca57\" (UID: \"54cce582-17c9-40d7-92f1-8a7c05a8ca57\") " Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.293754 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54cce582-17c9-40d7-92f1-8a7c05a8ca57-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "54cce582-17c9-40d7-92f1-8a7c05a8ca57" (UID: "54cce582-17c9-40d7-92f1-8a7c05a8ca57"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.293826 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwm6b\" (UniqueName: \"kubernetes.io/projected/48b7b556-b31c-4a94-aa25-3a26ba8d07dd-kube-api-access-fwm6b\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.298035 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54cce582-17c9-40d7-92f1-8a7c05a8ca57-kube-api-access-mdmpr" (OuterVolumeSpecName: "kube-api-access-mdmpr") pod "54cce582-17c9-40d7-92f1-8a7c05a8ca57" (UID: "54cce582-17c9-40d7-92f1-8a7c05a8ca57"). InnerVolumeSpecName "kube-api-access-mdmpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.395788 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54cce582-17c9-40d7-92f1-8a7c05a8ca57-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.395830 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdmpr\" (UniqueName: \"kubernetes.io/projected/54cce582-17c9-40d7-92f1-8a7c05a8ca57-kube-api-access-mdmpr\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.782756 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tzqff" event={"ID":"54cce582-17c9-40d7-92f1-8a7c05a8ca57","Type":"ContainerDied","Data":"9c52daac787d4422f6b69d8ad6a3c2d9831a9ca81550d36cb1d1eeac8f86dfee"} Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.782809 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c52daac787d4422f6b69d8ad6a3c2d9831a9ca81550d36cb1d1eeac8f86dfee" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.782816 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tzqff" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.786471 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0c633337-5e1d-438b-8b18-0acb62ec2da0","Type":"ContainerStarted","Data":"d414f34d90511b7735512304b7e2e0d05e454ca4546787b2f1fd13a79aaddd76"} Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.786629 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.790855 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1a670dab-0f15-4572-b2a8-36ec2330adb2","Type":"ContainerStarted","Data":"9c73bc96a9bd46a33bc825a170d9b9449a4f62874ff606906eaa78654b7bdebb"} Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.791834 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.796051 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-005f-account-create-update-d7fhm" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.796077 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-005f-account-create-update-d7fhm" event={"ID":"48b7b556-b31c-4a94-aa25-3a26ba8d07dd","Type":"ContainerDied","Data":"7cbe4f521a420f3411cb726f644db16f7f7ca3834b2411c4a030e058f41ab6f4"} Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.796157 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cbe4f521a420f3411cb726f644db16f7f7ca3834b2411c4a030e058f41ab6f4" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.829992 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.333047749 podStartE2EDuration="58.829976616s" podCreationTimestamp="2025-12-11 08:55:36 +0000 UTC" firstStartedPulling="2025-12-11 08:55:38.375082148 +0000 UTC m=+1126.440497756" lastFinishedPulling="2025-12-11 08:55:59.872011015 +0000 UTC m=+1147.937426623" observedRunningTime="2025-12-11 08:56:34.818803115 +0000 UTC m=+1182.884218713" watchObservedRunningTime="2025-12-11 08:56:34.829976616 +0000 UTC m=+1182.895392224" Dec 11 08:56:34 crc kubenswrapper[4629]: I1211 08:56:34.865994 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.177326059 podStartE2EDuration="59.861298159s" podCreationTimestamp="2025-12-11 08:55:35 +0000 UTC" firstStartedPulling="2025-12-11 08:55:38.195157298 +0000 UTC m=+1126.260572906" lastFinishedPulling="2025-12-11 08:55:59.879129398 +0000 UTC m=+1147.944545006" observedRunningTime="2025-12-11 08:56:34.847512426 +0000 UTC m=+1182.912928034" watchObservedRunningTime="2025-12-11 08:56:34.861298159 +0000 UTC m=+1182.926713767" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.242728 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a1e4-account-create-update-rlcdj" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.325444 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e81ddaf-9300-4a5b-b3cf-5f90928921a2-operator-scripts\") pod \"0e81ddaf-9300-4a5b-b3cf-5f90928921a2\" (UID: \"0e81ddaf-9300-4a5b-b3cf-5f90928921a2\") " Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.325920 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbwkg\" (UniqueName: \"kubernetes.io/projected/0e81ddaf-9300-4a5b-b3cf-5f90928921a2-kube-api-access-qbwkg\") pod \"0e81ddaf-9300-4a5b-b3cf-5f90928921a2\" (UID: \"0e81ddaf-9300-4a5b-b3cf-5f90928921a2\") " Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.327013 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e81ddaf-9300-4a5b-b3cf-5f90928921a2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e81ddaf-9300-4a5b-b3cf-5f90928921a2" (UID: "0e81ddaf-9300-4a5b-b3cf-5f90928921a2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.333269 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e81ddaf-9300-4a5b-b3cf-5f90928921a2-kube-api-access-qbwkg" (OuterVolumeSpecName: "kube-api-access-qbwkg") pod "0e81ddaf-9300-4a5b-b3cf-5f90928921a2" (UID: "0e81ddaf-9300-4a5b-b3cf-5f90928921a2"). InnerVolumeSpecName "kube-api-access-qbwkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.415205 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8495-account-create-update-5rsvg" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.432509 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbwkg\" (UniqueName: \"kubernetes.io/projected/0e81ddaf-9300-4a5b-b3cf-5f90928921a2-kube-api-access-qbwkg\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.432537 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e81ddaf-9300-4a5b-b3cf-5f90928921a2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.433997 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-grhlk" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.438122 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2sknt" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.533619 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/623e0d38-cbc4-4261-ad37-319eb1704ce2-operator-scripts\") pod \"623e0d38-cbc4-4261-ad37-319eb1704ce2\" (UID: \"623e0d38-cbc4-4261-ad37-319eb1704ce2\") " Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.533704 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drqzz\" (UniqueName: \"kubernetes.io/projected/623e0d38-cbc4-4261-ad37-319eb1704ce2-kube-api-access-drqzz\") pod \"623e0d38-cbc4-4261-ad37-319eb1704ce2\" (UID: \"623e0d38-cbc4-4261-ad37-319eb1704ce2\") " Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.533776 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac-operator-scripts\") pod \"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac\" (UID: \"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac\") " Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.533828 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btk2q\" (UniqueName: \"kubernetes.io/projected/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac-kube-api-access-btk2q\") pod \"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac\" (UID: \"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac\") " Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.533932 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cnp2\" (UniqueName: \"kubernetes.io/projected/3da40552-275b-4c71-a2d5-2257bdad664d-kube-api-access-7cnp2\") pod \"3da40552-275b-4c71-a2d5-2257bdad664d\" (UID: \"3da40552-275b-4c71-a2d5-2257bdad664d\") " Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.533961 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da40552-275b-4c71-a2d5-2257bdad664d-operator-scripts\") pod \"3da40552-275b-4c71-a2d5-2257bdad664d\" (UID: \"3da40552-275b-4c71-a2d5-2257bdad664d\") " Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.535400 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/623e0d38-cbc4-4261-ad37-319eb1704ce2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "623e0d38-cbc4-4261-ad37-319eb1704ce2" (UID: "623e0d38-cbc4-4261-ad37-319eb1704ce2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.536401 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b7f3b384-3bf7-46e1-9316-9c83bac9a8ac" (UID: "b7f3b384-3bf7-46e1-9316-9c83bac9a8ac"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.536672 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3da40552-275b-4c71-a2d5-2257bdad664d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3da40552-275b-4c71-a2d5-2257bdad664d" (UID: "3da40552-275b-4c71-a2d5-2257bdad664d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.540315 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/623e0d38-cbc4-4261-ad37-319eb1704ce2-kube-api-access-drqzz" (OuterVolumeSpecName: "kube-api-access-drqzz") pod "623e0d38-cbc4-4261-ad37-319eb1704ce2" (UID: "623e0d38-cbc4-4261-ad37-319eb1704ce2"). InnerVolumeSpecName "kube-api-access-drqzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.542554 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3da40552-275b-4c71-a2d5-2257bdad664d-kube-api-access-7cnp2" (OuterVolumeSpecName: "kube-api-access-7cnp2") pod "3da40552-275b-4c71-a2d5-2257bdad664d" (UID: "3da40552-275b-4c71-a2d5-2257bdad664d"). InnerVolumeSpecName "kube-api-access-7cnp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.542711 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac-kube-api-access-btk2q" (OuterVolumeSpecName: "kube-api-access-btk2q") pod "b7f3b384-3bf7-46e1-9316-9c83bac9a8ac" (UID: "b7f3b384-3bf7-46e1-9316-9c83bac9a8ac"). InnerVolumeSpecName "kube-api-access-btk2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.636372 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drqzz\" (UniqueName: \"kubernetes.io/projected/623e0d38-cbc4-4261-ad37-319eb1704ce2-kube-api-access-drqzz\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.636763 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.636863 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btk2q\" (UniqueName: \"kubernetes.io/projected/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac-kube-api-access-btk2q\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.636952 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cnp2\" (UniqueName: \"kubernetes.io/projected/3da40552-275b-4c71-a2d5-2257bdad664d-kube-api-access-7cnp2\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.637010 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da40552-275b-4c71-a2d5-2257bdad664d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.637071 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/623e0d38-cbc4-4261-ad37-319eb1704ce2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.804861 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2sknt" event={"ID":"3da40552-275b-4c71-a2d5-2257bdad664d","Type":"ContainerDied","Data":"785bd98d77ffeb6600d497ac0ba9f99e63b9efe2fc58e603d37acd2ceb67126e"} Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.804901 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="785bd98d77ffeb6600d497ac0ba9f99e63b9efe2fc58e603d37acd2ceb67126e" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.804966 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2sknt" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.810565 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8495-account-create-update-5rsvg" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.810582 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8495-account-create-update-5rsvg" event={"ID":"b7f3b384-3bf7-46e1-9316-9c83bac9a8ac","Type":"ContainerDied","Data":"79c5c82660811ea3cbbab07a33959038bdd5793bca634990137092bdb576147f"} Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.810625 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79c5c82660811ea3cbbab07a33959038bdd5793bca634990137092bdb576147f" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.812248 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a1e4-account-create-update-rlcdj" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.812239 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a1e4-account-create-update-rlcdj" event={"ID":"0e81ddaf-9300-4a5b-b3cf-5f90928921a2","Type":"ContainerDied","Data":"8d53c854c685da347b58cece36dc768c56351977cc5f234772e0591ba5540a67"} Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.812525 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d53c854c685da347b58cece36dc768c56351977cc5f234772e0591ba5540a67" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.813914 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-grhlk" event={"ID":"623e0d38-cbc4-4261-ad37-319eb1704ce2","Type":"ContainerDied","Data":"136b96aaf9d8f9e9022af40bc1c472e920980ad97288a1b568b2c7269c5a5ce6"} Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.813942 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="136b96aaf9d8f9e9022af40bc1c472e920980ad97288a1b568b2c7269c5a5ce6" Dec 11 08:56:35 crc kubenswrapper[4629]: I1211 08:56:35.814111 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-grhlk" Dec 11 08:56:36 crc kubenswrapper[4629]: I1211 08:56:36.188005 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.605627 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-j8tjt"] Dec 11 08:56:41 crc kubenswrapper[4629]: E1211 08:56:41.606914 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e81ddaf-9300-4a5b-b3cf-5f90928921a2" containerName="mariadb-account-create-update" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.606930 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e81ddaf-9300-4a5b-b3cf-5f90928921a2" containerName="mariadb-account-create-update" Dec 11 08:56:41 crc kubenswrapper[4629]: E1211 08:56:41.606941 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="623e0d38-cbc4-4261-ad37-319eb1704ce2" containerName="mariadb-database-create" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.606947 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="623e0d38-cbc4-4261-ad37-319eb1704ce2" containerName="mariadb-database-create" Dec 11 08:56:41 crc kubenswrapper[4629]: E1211 08:56:41.606954 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48b7b556-b31c-4a94-aa25-3a26ba8d07dd" containerName="mariadb-account-create-update" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.606963 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="48b7b556-b31c-4a94-aa25-3a26ba8d07dd" containerName="mariadb-account-create-update" Dec 11 08:56:41 crc kubenswrapper[4629]: E1211 08:56:41.606974 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da40552-275b-4c71-a2d5-2257bdad664d" containerName="mariadb-database-create" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.606980 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da40552-275b-4c71-a2d5-2257bdad664d" containerName="mariadb-database-create" Dec 11 08:56:41 crc kubenswrapper[4629]: E1211 08:56:41.606991 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f3b384-3bf7-46e1-9316-9c83bac9a8ac" containerName="mariadb-account-create-update" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.606996 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f3b384-3bf7-46e1-9316-9c83bac9a8ac" containerName="mariadb-account-create-update" Dec 11 08:56:41 crc kubenswrapper[4629]: E1211 08:56:41.607012 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cce582-17c9-40d7-92f1-8a7c05a8ca57" containerName="mariadb-database-create" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.607018 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cce582-17c9-40d7-92f1-8a7c05a8ca57" containerName="mariadb-database-create" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.607180 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e81ddaf-9300-4a5b-b3cf-5f90928921a2" containerName="mariadb-account-create-update" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.607200 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cce582-17c9-40d7-92f1-8a7c05a8ca57" containerName="mariadb-database-create" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.607209 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="623e0d38-cbc4-4261-ad37-319eb1704ce2" containerName="mariadb-database-create" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.607217 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7f3b384-3bf7-46e1-9316-9c83bac9a8ac" containerName="mariadb-account-create-update" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.607227 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="3da40552-275b-4c71-a2d5-2257bdad664d" containerName="mariadb-database-create" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.607237 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="48b7b556-b31c-4a94-aa25-3a26ba8d07dd" containerName="mariadb-account-create-update" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.607750 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.611601 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.612198 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-v482q" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.637913 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-j8tjt"] Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.783092 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-config-data\") pod \"glance-db-sync-j8tjt\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.783178 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tc7n\" (UniqueName: \"kubernetes.io/projected/c469e9bc-a78a-41cc-8034-1fbec2d2452e-kube-api-access-8tc7n\") pod \"glance-db-sync-j8tjt\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.783272 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-combined-ca-bundle\") pod \"glance-db-sync-j8tjt\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.783310 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-db-sync-config-data\") pod \"glance-db-sync-j8tjt\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.884343 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tc7n\" (UniqueName: \"kubernetes.io/projected/c469e9bc-a78a-41cc-8034-1fbec2d2452e-kube-api-access-8tc7n\") pod \"glance-db-sync-j8tjt\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.884404 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-combined-ca-bundle\") pod \"glance-db-sync-j8tjt\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.884452 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-db-sync-config-data\") pod \"glance-db-sync-j8tjt\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.884516 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-config-data\") pod \"glance-db-sync-j8tjt\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.895097 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-combined-ca-bundle\") pod \"glance-db-sync-j8tjt\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.895362 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-db-sync-config-data\") pod \"glance-db-sync-j8tjt\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.899436 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-config-data\") pod \"glance-db-sync-j8tjt\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.915496 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tc7n\" (UniqueName: \"kubernetes.io/projected/c469e9bc-a78a-41cc-8034-1fbec2d2452e-kube-api-access-8tc7n\") pod \"glance-db-sync-j8tjt\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:41 crc kubenswrapper[4629]: I1211 08:56:41.931268 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j8tjt" Dec 11 08:56:42 crc kubenswrapper[4629]: I1211 08:56:42.338928 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8lfjz" podUID="708ce743-d291-41e4-8c15-a9ad1cda240e" containerName="ovn-controller" probeResult="failure" output=< Dec 11 08:56:42 crc kubenswrapper[4629]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 11 08:56:42 crc kubenswrapper[4629]: > Dec 11 08:56:42 crc kubenswrapper[4629]: I1211 08:56:42.618456 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-j8tjt"] Dec 11 08:56:42 crc kubenswrapper[4629]: I1211 08:56:42.892640 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j8tjt" event={"ID":"c469e9bc-a78a-41cc-8034-1fbec2d2452e","Type":"ContainerStarted","Data":"01cb5f27edebfe7ca2bf195860c56f52149736b44807d0f0c8d186970788efe9"} Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.310764 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8lfjz" podUID="708ce743-d291-41e4-8c15-a9ad1cda240e" containerName="ovn-controller" probeResult="failure" output=< Dec 11 08:56:47 crc kubenswrapper[4629]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 11 08:56:47 crc kubenswrapper[4629]: > Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.407012 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.466418 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-v9ctv" Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.545043 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.671161 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.730499 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8lfjz-config-ftcvn"] Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.731786 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.738503 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.776406 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8lfjz-config-ftcvn"] Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.900923 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/613817ce-916e-4cb2-98cf-e55bd98c27f9-additional-scripts\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.900996 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-log-ovn\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.901062 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-run\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.901146 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-run-ovn\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.901165 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/613817ce-916e-4cb2-98cf-e55bd98c27f9-scripts\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:47 crc kubenswrapper[4629]: I1211 08:56:47.901201 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2lwf\" (UniqueName: \"kubernetes.io/projected/613817ce-916e-4cb2-98cf-e55bd98c27f9-kube-api-access-x2lwf\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.011168 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/613817ce-916e-4cb2-98cf-e55bd98c27f9-additional-scripts\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.011179 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/613817ce-916e-4cb2-98cf-e55bd98c27f9-additional-scripts\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.011577 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-log-ovn\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.011677 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-run\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.011758 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-run-ovn\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.011782 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/613817ce-916e-4cb2-98cf-e55bd98c27f9-scripts\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.011846 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2lwf\" (UniqueName: \"kubernetes.io/projected/613817ce-916e-4cb2-98cf-e55bd98c27f9-kube-api-access-x2lwf\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.012403 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-run\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.012514 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-run-ovn\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.012654 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-log-ovn\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.014532 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/613817ce-916e-4cb2-98cf-e55bd98c27f9-scripts\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.060022 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2lwf\" (UniqueName: \"kubernetes.io/projected/613817ce-916e-4cb2-98cf-e55bd98c27f9-kube-api-access-x2lwf\") pod \"ovn-controller-8lfjz-config-ftcvn\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.060458 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.404238 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-ww7bz"] Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.407668 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ww7bz" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.426938 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-ww7bz"] Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.519380 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-df7c-account-create-update-5f4c6"] Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.538956 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mptvf\" (UniqueName: \"kubernetes.io/projected/4ed3ed74-d821-4428-b47b-50fdc38d6bf2-kube-api-access-mptvf\") pod \"cinder-db-create-ww7bz\" (UID: \"4ed3ed74-d821-4428-b47b-50fdc38d6bf2\") " pod="openstack/cinder-db-create-ww7bz" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.539034 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ed3ed74-d821-4428-b47b-50fdc38d6bf2-operator-scripts\") pod \"cinder-db-create-ww7bz\" (UID: \"4ed3ed74-d821-4428-b47b-50fdc38d6bf2\") " pod="openstack/cinder-db-create-ww7bz" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.570899 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-df7c-account-create-update-5f4c6" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.571020 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-df7c-account-create-update-5f4c6"] Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.592799 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.620055 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-d2wsq"] Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.621287 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d2wsq" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.639931 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ed3ed74-d821-4428-b47b-50fdc38d6bf2-operator-scripts\") pod \"cinder-db-create-ww7bz\" (UID: \"4ed3ed74-d821-4428-b47b-50fdc38d6bf2\") " pod="openstack/cinder-db-create-ww7bz" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.640069 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mptvf\" (UniqueName: \"kubernetes.io/projected/4ed3ed74-d821-4428-b47b-50fdc38d6bf2-kube-api-access-mptvf\") pod \"cinder-db-create-ww7bz\" (UID: \"4ed3ed74-d821-4428-b47b-50fdc38d6bf2\") " pod="openstack/cinder-db-create-ww7bz" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.641142 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ed3ed74-d821-4428-b47b-50fdc38d6bf2-operator-scripts\") pod \"cinder-db-create-ww7bz\" (UID: \"4ed3ed74-d821-4428-b47b-50fdc38d6bf2\") " pod="openstack/cinder-db-create-ww7bz" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.686641 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-d2wsq"] Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.736608 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mptvf\" (UniqueName: \"kubernetes.io/projected/4ed3ed74-d821-4428-b47b-50fdc38d6bf2-kube-api-access-mptvf\") pod \"cinder-db-create-ww7bz\" (UID: \"4ed3ed74-d821-4428-b47b-50fdc38d6bf2\") " pod="openstack/cinder-db-create-ww7bz" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.741831 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b387bcfc-8cc5-42ac-a956-4d890b023403-operator-scripts\") pod \"barbican-db-create-d2wsq\" (UID: \"b387bcfc-8cc5-42ac-a956-4d890b023403\") " pod="openstack/barbican-db-create-d2wsq" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.741905 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71654a6c-0437-494a-9162-a1dcd1d3547f-operator-scripts\") pod \"barbican-df7c-account-create-update-5f4c6\" (UID: \"71654a6c-0437-494a-9162-a1dcd1d3547f\") " pod="openstack/barbican-df7c-account-create-update-5f4c6" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.741945 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5ncb\" (UniqueName: \"kubernetes.io/projected/71654a6c-0437-494a-9162-a1dcd1d3547f-kube-api-access-f5ncb\") pod \"barbican-df7c-account-create-update-5f4c6\" (UID: \"71654a6c-0437-494a-9162-a1dcd1d3547f\") " pod="openstack/barbican-df7c-account-create-update-5f4c6" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.741988 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmjwx\" (UniqueName: \"kubernetes.io/projected/b387bcfc-8cc5-42ac-a956-4d890b023403-kube-api-access-cmjwx\") pod \"barbican-db-create-d2wsq\" (UID: \"b387bcfc-8cc5-42ac-a956-4d890b023403\") " pod="openstack/barbican-db-create-d2wsq" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.796248 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ww7bz" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.847798 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b387bcfc-8cc5-42ac-a956-4d890b023403-operator-scripts\") pod \"barbican-db-create-d2wsq\" (UID: \"b387bcfc-8cc5-42ac-a956-4d890b023403\") " pod="openstack/barbican-db-create-d2wsq" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.847897 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71654a6c-0437-494a-9162-a1dcd1d3547f-operator-scripts\") pod \"barbican-df7c-account-create-update-5f4c6\" (UID: \"71654a6c-0437-494a-9162-a1dcd1d3547f\") " pod="openstack/barbican-df7c-account-create-update-5f4c6" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.847952 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5ncb\" (UniqueName: \"kubernetes.io/projected/71654a6c-0437-494a-9162-a1dcd1d3547f-kube-api-access-f5ncb\") pod \"barbican-df7c-account-create-update-5f4c6\" (UID: \"71654a6c-0437-494a-9162-a1dcd1d3547f\") " pod="openstack/barbican-df7c-account-create-update-5f4c6" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.847986 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmjwx\" (UniqueName: \"kubernetes.io/projected/b387bcfc-8cc5-42ac-a956-4d890b023403-kube-api-access-cmjwx\") pod \"barbican-db-create-d2wsq\" (UID: \"b387bcfc-8cc5-42ac-a956-4d890b023403\") " pod="openstack/barbican-db-create-d2wsq" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.849352 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71654a6c-0437-494a-9162-a1dcd1d3547f-operator-scripts\") pod \"barbican-df7c-account-create-update-5f4c6\" (UID: \"71654a6c-0437-494a-9162-a1dcd1d3547f\") " pod="openstack/barbican-df7c-account-create-update-5f4c6" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.850534 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b387bcfc-8cc5-42ac-a956-4d890b023403-operator-scripts\") pod \"barbican-db-create-d2wsq\" (UID: \"b387bcfc-8cc5-42ac-a956-4d890b023403\") " pod="openstack/barbican-db-create-d2wsq" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.886447 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-4a47-account-create-update-qbfpz"] Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.888123 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4a47-account-create-update-qbfpz" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.894479 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.987432 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmjwx\" (UniqueName: \"kubernetes.io/projected/b387bcfc-8cc5-42ac-a956-4d890b023403-kube-api-access-cmjwx\") pod \"barbican-db-create-d2wsq\" (UID: \"b387bcfc-8cc5-42ac-a956-4d890b023403\") " pod="openstack/barbican-db-create-d2wsq" Dec 11 08:56:48 crc kubenswrapper[4629]: I1211 08:56:48.990421 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5ncb\" (UniqueName: \"kubernetes.io/projected/71654a6c-0437-494a-9162-a1dcd1d3547f-kube-api-access-f5ncb\") pod \"barbican-df7c-account-create-update-5f4c6\" (UID: \"71654a6c-0437-494a-9162-a1dcd1d3547f\") " pod="openstack/barbican-df7c-account-create-update-5f4c6" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.021421 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8lfjz-config-ftcvn"] Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.034547 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-4a47-account-create-update-qbfpz"] Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.055078 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/221a021f-7c38-4604-b717-2a93af571c05-operator-scripts\") pod \"cinder-4a47-account-create-update-qbfpz\" (UID: \"221a021f-7c38-4604-b717-2a93af571c05\") " pod="openstack/cinder-4a47-account-create-update-qbfpz" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.055240 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rpg5\" (UniqueName: \"kubernetes.io/projected/221a021f-7c38-4604-b717-2a93af571c05-kube-api-access-9rpg5\") pod \"cinder-4a47-account-create-update-qbfpz\" (UID: \"221a021f-7c38-4604-b717-2a93af571c05\") " pod="openstack/cinder-4a47-account-create-update-qbfpz" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.071875 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d2wsq" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.156188 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/221a021f-7c38-4604-b717-2a93af571c05-operator-scripts\") pod \"cinder-4a47-account-create-update-qbfpz\" (UID: \"221a021f-7c38-4604-b717-2a93af571c05\") " pod="openstack/cinder-4a47-account-create-update-qbfpz" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.156355 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rpg5\" (UniqueName: \"kubernetes.io/projected/221a021f-7c38-4604-b717-2a93af571c05-kube-api-access-9rpg5\") pod \"cinder-4a47-account-create-update-qbfpz\" (UID: \"221a021f-7c38-4604-b717-2a93af571c05\") " pod="openstack/cinder-4a47-account-create-update-qbfpz" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.156923 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-fxcjb"] Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.157412 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/221a021f-7c38-4604-b717-2a93af571c05-operator-scripts\") pod \"cinder-4a47-account-create-update-qbfpz\" (UID: \"221a021f-7c38-4604-b717-2a93af571c05\") " pod="openstack/cinder-4a47-account-create-update-qbfpz" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.158409 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fxcjb" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.188456 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fxcjb"] Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.259003 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac490e45-b339-4803-b79e-075da951724b-operator-scripts\") pod \"neutron-db-create-fxcjb\" (UID: \"ac490e45-b339-4803-b79e-075da951724b\") " pod="openstack/neutron-db-create-fxcjb" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.259136 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcxm7\" (UniqueName: \"kubernetes.io/projected/ac490e45-b339-4803-b79e-075da951724b-kube-api-access-zcxm7\") pod \"neutron-db-create-fxcjb\" (UID: \"ac490e45-b339-4803-b79e-075da951724b\") " pod="openstack/neutron-db-create-fxcjb" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.262199 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-df7c-account-create-update-5f4c6" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.292699 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-gvk6g"] Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.293932 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.300478 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.301140 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.301488 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-m896j" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.301550 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rpg5\" (UniqueName: \"kubernetes.io/projected/221a021f-7c38-4604-b717-2a93af571c05-kube-api-access-9rpg5\") pod \"cinder-4a47-account-create-update-qbfpz\" (UID: \"221a021f-7c38-4604-b717-2a93af571c05\") " pod="openstack/cinder-4a47-account-create-update-qbfpz" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.363300 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac490e45-b339-4803-b79e-075da951724b-operator-scripts\") pod \"neutron-db-create-fxcjb\" (UID: \"ac490e45-b339-4803-b79e-075da951724b\") " pod="openstack/neutron-db-create-fxcjb" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.363427 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcxm7\" (UniqueName: \"kubernetes.io/projected/ac490e45-b339-4803-b79e-075da951724b-kube-api-access-zcxm7\") pod \"neutron-db-create-fxcjb\" (UID: \"ac490e45-b339-4803-b79e-075da951724b\") " pod="openstack/neutron-db-create-fxcjb" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.365135 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac490e45-b339-4803-b79e-075da951724b-operator-scripts\") pod \"neutron-db-create-fxcjb\" (UID: \"ac490e45-b339-4803-b79e-075da951724b\") " pod="openstack/neutron-db-create-fxcjb" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.466734 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pwlp\" (UniqueName: \"kubernetes.io/projected/29518fb5-bde7-4468-860f-37b425465f70-kube-api-access-8pwlp\") pod \"keystone-db-sync-gvk6g\" (UID: \"29518fb5-bde7-4468-860f-37b425465f70\") " pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.467197 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29518fb5-bde7-4468-860f-37b425465f70-config-data\") pod \"keystone-db-sync-gvk6g\" (UID: \"29518fb5-bde7-4468-860f-37b425465f70\") " pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.467296 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29518fb5-bde7-4468-860f-37b425465f70-combined-ca-bundle\") pod \"keystone-db-sync-gvk6g\" (UID: \"29518fb5-bde7-4468-860f-37b425465f70\") " pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.498530 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.523704 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gvk6g"] Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.551728 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4a47-account-create-update-qbfpz" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.569406 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pwlp\" (UniqueName: \"kubernetes.io/projected/29518fb5-bde7-4468-860f-37b425465f70-kube-api-access-8pwlp\") pod \"keystone-db-sync-gvk6g\" (UID: \"29518fb5-bde7-4468-860f-37b425465f70\") " pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.569445 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29518fb5-bde7-4468-860f-37b425465f70-config-data\") pod \"keystone-db-sync-gvk6g\" (UID: \"29518fb5-bde7-4468-860f-37b425465f70\") " pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.569521 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29518fb5-bde7-4468-860f-37b425465f70-combined-ca-bundle\") pod \"keystone-db-sync-gvk6g\" (UID: \"29518fb5-bde7-4468-860f-37b425465f70\") " pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.596677 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29518fb5-bde7-4468-860f-37b425465f70-combined-ca-bundle\") pod \"keystone-db-sync-gvk6g\" (UID: \"29518fb5-bde7-4468-860f-37b425465f70\") " pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.599482 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29518fb5-bde7-4468-860f-37b425465f70-config-data\") pod \"keystone-db-sync-gvk6g\" (UID: \"29518fb5-bde7-4468-860f-37b425465f70\") " pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.607530 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcxm7\" (UniqueName: \"kubernetes.io/projected/ac490e45-b339-4803-b79e-075da951724b-kube-api-access-zcxm7\") pod \"neutron-db-create-fxcjb\" (UID: \"ac490e45-b339-4803-b79e-075da951724b\") " pod="openstack/neutron-db-create-fxcjb" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.653907 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pwlp\" (UniqueName: \"kubernetes.io/projected/29518fb5-bde7-4468-860f-37b425465f70-kube-api-access-8pwlp\") pod \"keystone-db-sync-gvk6g\" (UID: \"29518fb5-bde7-4468-860f-37b425465f70\") " pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.669722 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-0b3c-account-create-update-7wnbl"] Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.765081 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0b3c-account-create-update-7wnbl" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.772134 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.785270 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fxcjb" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.806247 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-0b3c-account-create-update-7wnbl"] Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.860548 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.868623 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brtms\" (UniqueName: \"kubernetes.io/projected/161bc45b-0f32-488e-bda2-eb150afafcda-kube-api-access-brtms\") pod \"neutron-0b3c-account-create-update-7wnbl\" (UID: \"161bc45b-0f32-488e-bda2-eb150afafcda\") " pod="openstack/neutron-0b3c-account-create-update-7wnbl" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.877401 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/161bc45b-0f32-488e-bda2-eb150afafcda-operator-scripts\") pod \"neutron-0b3c-account-create-update-7wnbl\" (UID: \"161bc45b-0f32-488e-bda2-eb150afafcda\") " pod="openstack/neutron-0b3c-account-create-update-7wnbl" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.979190 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brtms\" (UniqueName: \"kubernetes.io/projected/161bc45b-0f32-488e-bda2-eb150afafcda-kube-api-access-brtms\") pod \"neutron-0b3c-account-create-update-7wnbl\" (UID: \"161bc45b-0f32-488e-bda2-eb150afafcda\") " pod="openstack/neutron-0b3c-account-create-update-7wnbl" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.979573 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/161bc45b-0f32-488e-bda2-eb150afafcda-operator-scripts\") pod \"neutron-0b3c-account-create-update-7wnbl\" (UID: \"161bc45b-0f32-488e-bda2-eb150afafcda\") " pod="openstack/neutron-0b3c-account-create-update-7wnbl" Dec 11 08:56:49 crc kubenswrapper[4629]: I1211 08:56:49.980727 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/161bc45b-0f32-488e-bda2-eb150afafcda-operator-scripts\") pod \"neutron-0b3c-account-create-update-7wnbl\" (UID: \"161bc45b-0f32-488e-bda2-eb150afafcda\") " pod="openstack/neutron-0b3c-account-create-update-7wnbl" Dec 11 08:56:50 crc kubenswrapper[4629]: I1211 08:56:50.039332 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brtms\" (UniqueName: \"kubernetes.io/projected/161bc45b-0f32-488e-bda2-eb150afafcda-kube-api-access-brtms\") pod \"neutron-0b3c-account-create-update-7wnbl\" (UID: \"161bc45b-0f32-488e-bda2-eb150afafcda\") " pod="openstack/neutron-0b3c-account-create-update-7wnbl" Dec 11 08:56:50 crc kubenswrapper[4629]: I1211 08:56:50.051554 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8lfjz-config-ftcvn" event={"ID":"613817ce-916e-4cb2-98cf-e55bd98c27f9","Type":"ContainerStarted","Data":"677ec427144b329476c2c355f3860ddb959651d40af1e4d03cd98b11770d8f36"} Dec 11 08:56:50 crc kubenswrapper[4629]: I1211 08:56:50.160995 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0b3c-account-create-update-7wnbl" Dec 11 08:56:50 crc kubenswrapper[4629]: I1211 08:56:50.486006 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:56:50 crc kubenswrapper[4629]: I1211 08:56:50.486329 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:56:50 crc kubenswrapper[4629]: I1211 08:56:50.486395 4629 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 08:56:50 crc kubenswrapper[4629]: I1211 08:56:50.488357 4629 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"921d43c980dda9b89debb49d73c6ac3dda5f931588d36bb68efdef51a59b46b8"} pod="openshift-machine-config-operator/machine-config-daemon-wbld9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:56:50 crc kubenswrapper[4629]: I1211 08:56:50.488445 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" containerID="cri-o://921d43c980dda9b89debb49d73c6ac3dda5f931588d36bb68efdef51a59b46b8" gracePeriod=600 Dec 11 08:56:50 crc kubenswrapper[4629]: I1211 08:56:50.547136 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="762352f0-3934-4d44-b846-fa922ffedd1d" containerName="galera" probeResult="failure" output="command timed out" Dec 11 08:56:50 crc kubenswrapper[4629]: I1211 08:56:50.599385 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-ww7bz"] Dec 11 08:56:50 crc kubenswrapper[4629]: E1211 08:56:50.810642 4629 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.111:52008->38.129.56.111:44939: write tcp 38.129.56.111:52008->38.129.56.111:44939: write: broken pipe Dec 11 08:56:50 crc kubenswrapper[4629]: I1211 08:56:50.941531 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-4a47-account-create-update-qbfpz"] Dec 11 08:56:50 crc kubenswrapper[4629]: I1211 08:56:50.982824 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-d2wsq"] Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.001435 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-df7c-account-create-update-5f4c6"] Dec 11 08:56:51 crc kubenswrapper[4629]: W1211 08:56:51.024129 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod221a021f_7c38_4604_b717_2a93af571c05.slice/crio-9d323c78ff596cf310893520607cf68a699555742953389d1d156e5d3337a4a9 WatchSource:0}: Error finding container 9d323c78ff596cf310893520607cf68a699555742953389d1d156e5d3337a4a9: Status 404 returned error can't find the container with id 9d323c78ff596cf310893520607cf68a699555742953389d1d156e5d3337a4a9 Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.089156 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ww7bz" event={"ID":"4ed3ed74-d821-4428-b47b-50fdc38d6bf2","Type":"ContainerStarted","Data":"7add405638732923a6971035747e3d371f9a62f8e70cabf2ca58ecd58bc7a1dc"} Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.113901 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8lfjz-config-ftcvn" event={"ID":"613817ce-916e-4cb2-98cf-e55bd98c27f9","Type":"ContainerStarted","Data":"ddea03484aeae20a8f71069e2fc25103d6b621f967af7e32be81c909767b96a5"} Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.127863 4629 generic.go:334] "Generic (PLEG): container finished" podID="847b50b9-0093-404f-b0dd-63c814d8986b" containerID="921d43c980dda9b89debb49d73c6ac3dda5f931588d36bb68efdef51a59b46b8" exitCode=0 Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.127993 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerDied","Data":"921d43c980dda9b89debb49d73c6ac3dda5f931588d36bb68efdef51a59b46b8"} Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.128031 4629 scope.go:117] "RemoveContainer" containerID="3478d1305c4015a5af7c390010bcd373d9a5066e2359bd5ed303da8f30af223b" Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.144579 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-0b3c-account-create-update-7wnbl"] Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.164022 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d2wsq" event={"ID":"b387bcfc-8cc5-42ac-a956-4d890b023403","Type":"ContainerStarted","Data":"41736078cad50daa0ef0a1d7d69b0652b8bf0cbf2e5baa3e8bc288e2f20485f6"} Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.174154 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fxcjb"] Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.174502 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-df7c-account-create-update-5f4c6" event={"ID":"71654a6c-0437-494a-9162-a1dcd1d3547f","Type":"ContainerStarted","Data":"0138d3becce33181ea3a443171e78cb3d5c4476c6131353dbfebe1ae698bbe3d"} Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.175473 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4a47-account-create-update-qbfpz" event={"ID":"221a021f-7c38-4604-b717-2a93af571c05","Type":"ContainerStarted","Data":"9d323c78ff596cf310893520607cf68a699555742953389d1d156e5d3337a4a9"} Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.196806 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gvk6g"] Dec 11 08:56:51 crc kubenswrapper[4629]: I1211 08:56:51.205002 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-8lfjz-config-ftcvn" podStartSLOduration=4.20498329 podStartE2EDuration="4.20498329s" podCreationTimestamp="2025-12-11 08:56:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:56:51.187243813 +0000 UTC m=+1199.252659441" watchObservedRunningTime="2025-12-11 08:56:51.20498329 +0000 UTC m=+1199.270398888" Dec 11 08:56:51 crc kubenswrapper[4629]: W1211 08:56:51.224539 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac490e45_b339_4803_b79e_075da951724b.slice/crio-122e4fa4a443359956134a71015207bc9567e0f47cad6504dad2510259f79f65 WatchSource:0}: Error finding container 122e4fa4a443359956134a71015207bc9567e0f47cad6504dad2510259f79f65: Status 404 returned error can't find the container with id 122e4fa4a443359956134a71015207bc9567e0f47cad6504dad2510259f79f65 Dec 11 08:56:52 crc kubenswrapper[4629]: I1211 08:56:52.213621 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4a47-account-create-update-qbfpz" event={"ID":"221a021f-7c38-4604-b717-2a93af571c05","Type":"ContainerStarted","Data":"941118659c828ad5c0454bbf7f8f880df80ff439fed7cbc1eaa72e36101fdb27"} Dec 11 08:56:52 crc kubenswrapper[4629]: I1211 08:56:52.215543 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gvk6g" event={"ID":"29518fb5-bde7-4468-860f-37b425465f70","Type":"ContainerStarted","Data":"c57e859bbe5feda781c486fbd775bef4aa5cfaecdf022fcba78dd0154bc5d7e8"} Dec 11 08:56:52 crc kubenswrapper[4629]: I1211 08:56:52.229947 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0b3c-account-create-update-7wnbl" event={"ID":"161bc45b-0f32-488e-bda2-eb150afafcda","Type":"ContainerStarted","Data":"33191f756e44007c10dd343fdc8a15bae8804ea231ab780e26b32cd0ba26036f"} Dec 11 08:56:52 crc kubenswrapper[4629]: I1211 08:56:52.235379 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fxcjb" event={"ID":"ac490e45-b339-4803-b79e-075da951724b","Type":"ContainerStarted","Data":"122e4fa4a443359956134a71015207bc9567e0f47cad6504dad2510259f79f65"} Dec 11 08:56:52 crc kubenswrapper[4629]: I1211 08:56:52.253608 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ww7bz" event={"ID":"4ed3ed74-d821-4428-b47b-50fdc38d6bf2","Type":"ContainerStarted","Data":"40645b8724f7042c62ffe3075af810dd6c88b2a6eb11e7dbc43ec1b2880b318f"} Dec 11 08:56:52 crc kubenswrapper[4629]: I1211 08:56:52.749518 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-8lfjz" Dec 11 08:56:52 crc kubenswrapper[4629]: I1211 08:56:52.794651 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-4a47-account-create-update-qbfpz" podStartSLOduration=4.794628615 podStartE2EDuration="4.794628615s" podCreationTimestamp="2025-12-11 08:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:56:52.736458739 +0000 UTC m=+1200.801874337" watchObservedRunningTime="2025-12-11 08:56:52.794628615 +0000 UTC m=+1200.860044223" Dec 11 08:56:52 crc kubenswrapper[4629]: I1211 08:56:52.804085 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-ww7bz" podStartSLOduration=4.804064411 podStartE2EDuration="4.804064411s" podCreationTimestamp="2025-12-11 08:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:56:52.795311606 +0000 UTC m=+1200.860727204" watchObservedRunningTime="2025-12-11 08:56:52.804064411 +0000 UTC m=+1200.869480019" Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.305645 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0b3c-account-create-update-7wnbl" event={"ID":"161bc45b-0f32-488e-bda2-eb150afafcda","Type":"ContainerStarted","Data":"98634e21d056b976e6f7d6b62c7ffa718ce74be85a1573ba4491002e5e5a3eb7"} Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.315802 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fxcjb" event={"ID":"ac490e45-b339-4803-b79e-075da951724b","Type":"ContainerStarted","Data":"643e2f4f71712537960ccc40d9b0caa65a1eb2cfcf109ca3fbb44aa1143b6efb"} Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.333932 4629 generic.go:334] "Generic (PLEG): container finished" podID="4ed3ed74-d821-4428-b47b-50fdc38d6bf2" containerID="40645b8724f7042c62ffe3075af810dd6c88b2a6eb11e7dbc43ec1b2880b318f" exitCode=0 Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.334017 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ww7bz" event={"ID":"4ed3ed74-d821-4428-b47b-50fdc38d6bf2","Type":"ContainerDied","Data":"40645b8724f7042c62ffe3075af810dd6c88b2a6eb11e7dbc43ec1b2880b318f"} Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.344835 4629 generic.go:334] "Generic (PLEG): container finished" podID="613817ce-916e-4cb2-98cf-e55bd98c27f9" containerID="ddea03484aeae20a8f71069e2fc25103d6b621f967af7e32be81c909767b96a5" exitCode=0 Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.345225 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8lfjz-config-ftcvn" event={"ID":"613817ce-916e-4cb2-98cf-e55bd98c27f9","Type":"ContainerDied","Data":"ddea03484aeae20a8f71069e2fc25103d6b621f967af7e32be81c909767b96a5"} Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.355782 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-0b3c-account-create-update-7wnbl" podStartSLOduration=4.355758428 podStartE2EDuration="4.355758428s" podCreationTimestamp="2025-12-11 08:56:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:56:53.354271181 +0000 UTC m=+1201.419686789" watchObservedRunningTime="2025-12-11 08:56:53.355758428 +0000 UTC m=+1201.421174036" Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.355898 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"f34698209a804ac13c8c04a3199ac3381f69055cef9be6fa18f868fa21af6591"} Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.364693 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d2wsq" event={"ID":"b387bcfc-8cc5-42ac-a956-4d890b023403","Type":"ContainerStarted","Data":"04432ad6ca030050521700b021ec0a7a68bd5c5b0ce9bc42f01293244144941a"} Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.368903 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-df7c-account-create-update-5f4c6" event={"ID":"71654a6c-0437-494a-9162-a1dcd1d3547f","Type":"ContainerStarted","Data":"6494e783f8ad29f5503fa3c93dd38053454239eb9cf7e0d97c1c237f3efab794"} Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.433277 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-fxcjb" podStartSLOduration=4.43325604 podStartE2EDuration="4.43325604s" podCreationTimestamp="2025-12-11 08:56:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:56:53.430322298 +0000 UTC m=+1201.495737896" watchObservedRunningTime="2025-12-11 08:56:53.43325604 +0000 UTC m=+1201.498671648" Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.623377 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-df7c-account-create-update-5f4c6" podStartSLOduration=5.623356447 podStartE2EDuration="5.623356447s" podCreationTimestamp="2025-12-11 08:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:56:53.61995157 +0000 UTC m=+1201.685367198" watchObservedRunningTime="2025-12-11 08:56:53.623356447 +0000 UTC m=+1201.688772055" Dec 11 08:56:53 crc kubenswrapper[4629]: I1211 08:56:53.689021 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-d2wsq" podStartSLOduration=5.6889988670000005 podStartE2EDuration="5.688998867s" podCreationTimestamp="2025-12-11 08:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:56:53.685639822 +0000 UTC m=+1201.751055430" watchObservedRunningTime="2025-12-11 08:56:53.688998867 +0000 UTC m=+1201.754414475" Dec 11 08:56:54 crc kubenswrapper[4629]: I1211 08:56:54.388254 4629 generic.go:334] "Generic (PLEG): container finished" podID="71654a6c-0437-494a-9162-a1dcd1d3547f" containerID="6494e783f8ad29f5503fa3c93dd38053454239eb9cf7e0d97c1c237f3efab794" exitCode=0 Dec 11 08:56:54 crc kubenswrapper[4629]: I1211 08:56:54.388368 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-df7c-account-create-update-5f4c6" event={"ID":"71654a6c-0437-494a-9162-a1dcd1d3547f","Type":"ContainerDied","Data":"6494e783f8ad29f5503fa3c93dd38053454239eb9cf7e0d97c1c237f3efab794"} Dec 11 08:56:54 crc kubenswrapper[4629]: I1211 08:56:54.396621 4629 generic.go:334] "Generic (PLEG): container finished" podID="221a021f-7c38-4604-b717-2a93af571c05" containerID="941118659c828ad5c0454bbf7f8f880df80ff439fed7cbc1eaa72e36101fdb27" exitCode=0 Dec 11 08:56:54 crc kubenswrapper[4629]: I1211 08:56:54.396709 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4a47-account-create-update-qbfpz" event={"ID":"221a021f-7c38-4604-b717-2a93af571c05","Type":"ContainerDied","Data":"941118659c828ad5c0454bbf7f8f880df80ff439fed7cbc1eaa72e36101fdb27"} Dec 11 08:56:54 crc kubenswrapper[4629]: I1211 08:56:54.411794 4629 generic.go:334] "Generic (PLEG): container finished" podID="161bc45b-0f32-488e-bda2-eb150afafcda" containerID="98634e21d056b976e6f7d6b62c7ffa718ce74be85a1573ba4491002e5e5a3eb7" exitCode=0 Dec 11 08:56:54 crc kubenswrapper[4629]: I1211 08:56:54.411892 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0b3c-account-create-update-7wnbl" event={"ID":"161bc45b-0f32-488e-bda2-eb150afafcda","Type":"ContainerDied","Data":"98634e21d056b976e6f7d6b62c7ffa718ce74be85a1573ba4491002e5e5a3eb7"} Dec 11 08:56:54 crc kubenswrapper[4629]: I1211 08:56:54.413541 4629 generic.go:334] "Generic (PLEG): container finished" podID="ac490e45-b339-4803-b79e-075da951724b" containerID="643e2f4f71712537960ccc40d9b0caa65a1eb2cfcf109ca3fbb44aa1143b6efb" exitCode=0 Dec 11 08:56:54 crc kubenswrapper[4629]: I1211 08:56:54.413588 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fxcjb" event={"ID":"ac490e45-b339-4803-b79e-075da951724b","Type":"ContainerDied","Data":"643e2f4f71712537960ccc40d9b0caa65a1eb2cfcf109ca3fbb44aa1143b6efb"} Dec 11 08:56:54 crc kubenswrapper[4629]: I1211 08:56:54.415370 4629 generic.go:334] "Generic (PLEG): container finished" podID="b387bcfc-8cc5-42ac-a956-4d890b023403" containerID="04432ad6ca030050521700b021ec0a7a68bd5c5b0ce9bc42f01293244144941a" exitCode=0 Dec 11 08:56:54 crc kubenswrapper[4629]: I1211 08:56:54.415509 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d2wsq" event={"ID":"b387bcfc-8cc5-42ac-a956-4d890b023403","Type":"ContainerDied","Data":"04432ad6ca030050521700b021ec0a7a68bd5c5b0ce9bc42f01293244144941a"} Dec 11 08:57:04 crc kubenswrapper[4629]: E1211 08:57:04.236176 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 11 08:57:04 crc kubenswrapper[4629]: E1211 08:57:04.236884 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8tc7n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-j8tjt_openstack(c469e9bc-a78a-41cc-8034-1fbec2d2452e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:57:04 crc kubenswrapper[4629]: E1211 08:57:04.238081 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-j8tjt" podUID="c469e9bc-a78a-41cc-8034-1fbec2d2452e" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.284630 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fxcjb" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.308198 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.323498 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ww7bz" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.364831 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-run\") pod \"613817ce-916e-4cb2-98cf-e55bd98c27f9\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.364922 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcxm7\" (UniqueName: \"kubernetes.io/projected/ac490e45-b339-4803-b79e-075da951724b-kube-api-access-zcxm7\") pod \"ac490e45-b339-4803-b79e-075da951724b\" (UID: \"ac490e45-b339-4803-b79e-075da951724b\") " Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.364927 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-run" (OuterVolumeSpecName: "var-run") pod "613817ce-916e-4cb2-98cf-e55bd98c27f9" (UID: "613817ce-916e-4cb2-98cf-e55bd98c27f9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.364964 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ed3ed74-d821-4428-b47b-50fdc38d6bf2-operator-scripts\") pod \"4ed3ed74-d821-4428-b47b-50fdc38d6bf2\" (UID: \"4ed3ed74-d821-4428-b47b-50fdc38d6bf2\") " Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.364992 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-log-ovn\") pod \"613817ce-916e-4cb2-98cf-e55bd98c27f9\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.365029 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/613817ce-916e-4cb2-98cf-e55bd98c27f9-additional-scripts\") pod \"613817ce-916e-4cb2-98cf-e55bd98c27f9\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.365076 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2lwf\" (UniqueName: \"kubernetes.io/projected/613817ce-916e-4cb2-98cf-e55bd98c27f9-kube-api-access-x2lwf\") pod \"613817ce-916e-4cb2-98cf-e55bd98c27f9\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.365099 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/613817ce-916e-4cb2-98cf-e55bd98c27f9-scripts\") pod \"613817ce-916e-4cb2-98cf-e55bd98c27f9\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.365119 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mptvf\" (UniqueName: \"kubernetes.io/projected/4ed3ed74-d821-4428-b47b-50fdc38d6bf2-kube-api-access-mptvf\") pod \"4ed3ed74-d821-4428-b47b-50fdc38d6bf2\" (UID: \"4ed3ed74-d821-4428-b47b-50fdc38d6bf2\") " Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.365182 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-run-ovn\") pod \"613817ce-916e-4cb2-98cf-e55bd98c27f9\" (UID: \"613817ce-916e-4cb2-98cf-e55bd98c27f9\") " Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.365234 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac490e45-b339-4803-b79e-075da951724b-operator-scripts\") pod \"ac490e45-b339-4803-b79e-075da951724b\" (UID: \"ac490e45-b339-4803-b79e-075da951724b\") " Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.365979 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "613817ce-916e-4cb2-98cf-e55bd98c27f9" (UID: "613817ce-916e-4cb2-98cf-e55bd98c27f9"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.366275 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "613817ce-916e-4cb2-98cf-e55bd98c27f9" (UID: "613817ce-916e-4cb2-98cf-e55bd98c27f9"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.366632 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ed3ed74-d821-4428-b47b-50fdc38d6bf2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ed3ed74-d821-4428-b47b-50fdc38d6bf2" (UID: "4ed3ed74-d821-4428-b47b-50fdc38d6bf2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.366979 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/613817ce-916e-4cb2-98cf-e55bd98c27f9-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "613817ce-916e-4cb2-98cf-e55bd98c27f9" (UID: "613817ce-916e-4cb2-98cf-e55bd98c27f9"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.367278 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/613817ce-916e-4cb2-98cf-e55bd98c27f9-scripts" (OuterVolumeSpecName: "scripts") pod "613817ce-916e-4cb2-98cf-e55bd98c27f9" (UID: "613817ce-916e-4cb2-98cf-e55bd98c27f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.367839 4629 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-run\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.367884 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ed3ed74-d821-4428-b47b-50fdc38d6bf2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.367900 4629 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.367915 4629 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/613817ce-916e-4cb2-98cf-e55bd98c27f9-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.367925 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/613817ce-916e-4cb2-98cf-e55bd98c27f9-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.367936 4629 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/613817ce-916e-4cb2-98cf-e55bd98c27f9-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.368975 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac490e45-b339-4803-b79e-075da951724b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ac490e45-b339-4803-b79e-075da951724b" (UID: "ac490e45-b339-4803-b79e-075da951724b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.370247 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac490e45-b339-4803-b79e-075da951724b-kube-api-access-zcxm7" (OuterVolumeSpecName: "kube-api-access-zcxm7") pod "ac490e45-b339-4803-b79e-075da951724b" (UID: "ac490e45-b339-4803-b79e-075da951724b"). InnerVolumeSpecName "kube-api-access-zcxm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.371129 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/613817ce-916e-4cb2-98cf-e55bd98c27f9-kube-api-access-x2lwf" (OuterVolumeSpecName: "kube-api-access-x2lwf") pod "613817ce-916e-4cb2-98cf-e55bd98c27f9" (UID: "613817ce-916e-4cb2-98cf-e55bd98c27f9"). InnerVolumeSpecName "kube-api-access-x2lwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.374044 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ed3ed74-d821-4428-b47b-50fdc38d6bf2-kube-api-access-mptvf" (OuterVolumeSpecName: "kube-api-access-mptvf") pod "4ed3ed74-d821-4428-b47b-50fdc38d6bf2" (UID: "4ed3ed74-d821-4428-b47b-50fdc38d6bf2"). InnerVolumeSpecName "kube-api-access-mptvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.470104 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcxm7\" (UniqueName: \"kubernetes.io/projected/ac490e45-b339-4803-b79e-075da951724b-kube-api-access-zcxm7\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.470148 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2lwf\" (UniqueName: \"kubernetes.io/projected/613817ce-916e-4cb2-98cf-e55bd98c27f9-kube-api-access-x2lwf\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.470160 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mptvf\" (UniqueName: \"kubernetes.io/projected/4ed3ed74-d821-4428-b47b-50fdc38d6bf2-kube-api-access-mptvf\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.470172 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac490e45-b339-4803-b79e-075da951724b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.519298 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fxcjb" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.519302 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fxcjb" event={"ID":"ac490e45-b339-4803-b79e-075da951724b","Type":"ContainerDied","Data":"122e4fa4a443359956134a71015207bc9567e0f47cad6504dad2510259f79f65"} Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.519415 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="122e4fa4a443359956134a71015207bc9567e0f47cad6504dad2510259f79f65" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.521228 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ww7bz" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.521218 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ww7bz" event={"ID":"4ed3ed74-d821-4428-b47b-50fdc38d6bf2","Type":"ContainerDied","Data":"7add405638732923a6971035747e3d371f9a62f8e70cabf2ca58ecd58bc7a1dc"} Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.521370 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7add405638732923a6971035747e3d371f9a62f8e70cabf2ca58ecd58bc7a1dc" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.523881 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8lfjz-config-ftcvn" Dec 11 08:57:04 crc kubenswrapper[4629]: E1211 08:57:04.525826 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-j8tjt" podUID="c469e9bc-a78a-41cc-8034-1fbec2d2452e" Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.526076 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8lfjz-config-ftcvn" event={"ID":"613817ce-916e-4cb2-98cf-e55bd98c27f9","Type":"ContainerDied","Data":"677ec427144b329476c2c355f3860ddb959651d40af1e4d03cd98b11770d8f36"} Dec 11 08:57:04 crc kubenswrapper[4629]: I1211 08:57:04.526114 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="677ec427144b329476c2c355f3860ddb959651d40af1e4d03cd98b11770d8f36" Dec 11 08:57:05 crc kubenswrapper[4629]: I1211 08:57:05.444282 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8lfjz-config-ftcvn"] Dec 11 08:57:05 crc kubenswrapper[4629]: I1211 08:57:05.462624 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8lfjz-config-ftcvn"] Dec 11 08:57:06 crc kubenswrapper[4629]: I1211 08:57:06.212321 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="613817ce-916e-4cb2-98cf-e55bd98c27f9" path="/var/lib/kubelet/pods/613817ce-916e-4cb2-98cf-e55bd98c27f9/volumes" Dec 11 08:57:06 crc kubenswrapper[4629]: E1211 08:57:06.841962 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-keystone:current-podified" Dec 11 08:57:06 crc kubenswrapper[4629]: E1211 08:57:06.842189 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:keystone-db-sync,Image:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,Command:[/bin/bash],Args:[-c keystone-manage db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/keystone/keystone.conf,SubPath:keystone.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8pwlp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42425,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42425,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-db-sync-gvk6g_openstack(29518fb5-bde7-4468-860f-37b425465f70): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:57:06 crc kubenswrapper[4629]: E1211 08:57:06.844384 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/keystone-db-sync-gvk6g" podUID="29518fb5-bde7-4468-860f-37b425465f70" Dec 11 08:57:06 crc kubenswrapper[4629]: I1211 08:57:06.953191 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d2wsq" Dec 11 08:57:06 crc kubenswrapper[4629]: I1211 08:57:06.960879 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4a47-account-create-update-qbfpz" Dec 11 08:57:06 crc kubenswrapper[4629]: I1211 08:57:06.978998 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-df7c-account-create-update-5f4c6" Dec 11 08:57:06 crc kubenswrapper[4629]: I1211 08:57:06.982897 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0b3c-account-create-update-7wnbl" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.021873 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5ncb\" (UniqueName: \"kubernetes.io/projected/71654a6c-0437-494a-9162-a1dcd1d3547f-kube-api-access-f5ncb\") pod \"71654a6c-0437-494a-9162-a1dcd1d3547f\" (UID: \"71654a6c-0437-494a-9162-a1dcd1d3547f\") " Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.021927 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b387bcfc-8cc5-42ac-a956-4d890b023403-operator-scripts\") pod \"b387bcfc-8cc5-42ac-a956-4d890b023403\" (UID: \"b387bcfc-8cc5-42ac-a956-4d890b023403\") " Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.021976 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rpg5\" (UniqueName: \"kubernetes.io/projected/221a021f-7c38-4604-b717-2a93af571c05-kube-api-access-9rpg5\") pod \"221a021f-7c38-4604-b717-2a93af571c05\" (UID: \"221a021f-7c38-4604-b717-2a93af571c05\") " Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.021996 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/221a021f-7c38-4604-b717-2a93af571c05-operator-scripts\") pod \"221a021f-7c38-4604-b717-2a93af571c05\" (UID: \"221a021f-7c38-4604-b717-2a93af571c05\") " Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.022071 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/161bc45b-0f32-488e-bda2-eb150afafcda-operator-scripts\") pod \"161bc45b-0f32-488e-bda2-eb150afafcda\" (UID: \"161bc45b-0f32-488e-bda2-eb150afafcda\") " Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.022119 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71654a6c-0437-494a-9162-a1dcd1d3547f-operator-scripts\") pod \"71654a6c-0437-494a-9162-a1dcd1d3547f\" (UID: \"71654a6c-0437-494a-9162-a1dcd1d3547f\") " Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.022147 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brtms\" (UniqueName: \"kubernetes.io/projected/161bc45b-0f32-488e-bda2-eb150afafcda-kube-api-access-brtms\") pod \"161bc45b-0f32-488e-bda2-eb150afafcda\" (UID: \"161bc45b-0f32-488e-bda2-eb150afafcda\") " Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.022166 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmjwx\" (UniqueName: \"kubernetes.io/projected/b387bcfc-8cc5-42ac-a956-4d890b023403-kube-api-access-cmjwx\") pod \"b387bcfc-8cc5-42ac-a956-4d890b023403\" (UID: \"b387bcfc-8cc5-42ac-a956-4d890b023403\") " Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.023188 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b387bcfc-8cc5-42ac-a956-4d890b023403-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b387bcfc-8cc5-42ac-a956-4d890b023403" (UID: "b387bcfc-8cc5-42ac-a956-4d890b023403"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.023649 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71654a6c-0437-494a-9162-a1dcd1d3547f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "71654a6c-0437-494a-9162-a1dcd1d3547f" (UID: "71654a6c-0437-494a-9162-a1dcd1d3547f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.026133 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/161bc45b-0f32-488e-bda2-eb150afafcda-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "161bc45b-0f32-488e-bda2-eb150afafcda" (UID: "161bc45b-0f32-488e-bda2-eb150afafcda"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.026956 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/221a021f-7c38-4604-b717-2a93af571c05-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "221a021f-7c38-4604-b717-2a93af571c05" (UID: "221a021f-7c38-4604-b717-2a93af571c05"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.031123 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71654a6c-0437-494a-9162-a1dcd1d3547f-kube-api-access-f5ncb" (OuterVolumeSpecName: "kube-api-access-f5ncb") pod "71654a6c-0437-494a-9162-a1dcd1d3547f" (UID: "71654a6c-0437-494a-9162-a1dcd1d3547f"). InnerVolumeSpecName "kube-api-access-f5ncb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.040792 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b387bcfc-8cc5-42ac-a956-4d890b023403-kube-api-access-cmjwx" (OuterVolumeSpecName: "kube-api-access-cmjwx") pod "b387bcfc-8cc5-42ac-a956-4d890b023403" (UID: "b387bcfc-8cc5-42ac-a956-4d890b023403"). InnerVolumeSpecName "kube-api-access-cmjwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.044187 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/221a021f-7c38-4604-b717-2a93af571c05-kube-api-access-9rpg5" (OuterVolumeSpecName: "kube-api-access-9rpg5") pod "221a021f-7c38-4604-b717-2a93af571c05" (UID: "221a021f-7c38-4604-b717-2a93af571c05"). InnerVolumeSpecName "kube-api-access-9rpg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.044314 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/161bc45b-0f32-488e-bda2-eb150afafcda-kube-api-access-brtms" (OuterVolumeSpecName: "kube-api-access-brtms") pod "161bc45b-0f32-488e-bda2-eb150afafcda" (UID: "161bc45b-0f32-488e-bda2-eb150afafcda"). InnerVolumeSpecName "kube-api-access-brtms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.123679 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rpg5\" (UniqueName: \"kubernetes.io/projected/221a021f-7c38-4604-b717-2a93af571c05-kube-api-access-9rpg5\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.123718 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/221a021f-7c38-4604-b717-2a93af571c05-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.123727 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/161bc45b-0f32-488e-bda2-eb150afafcda-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.123736 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71654a6c-0437-494a-9162-a1dcd1d3547f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.123744 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brtms\" (UniqueName: \"kubernetes.io/projected/161bc45b-0f32-488e-bda2-eb150afafcda-kube-api-access-brtms\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.123755 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmjwx\" (UniqueName: \"kubernetes.io/projected/b387bcfc-8cc5-42ac-a956-4d890b023403-kube-api-access-cmjwx\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.123763 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5ncb\" (UniqueName: \"kubernetes.io/projected/71654a6c-0437-494a-9162-a1dcd1d3547f-kube-api-access-f5ncb\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.123772 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b387bcfc-8cc5-42ac-a956-4d890b023403-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.564923 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d2wsq" event={"ID":"b387bcfc-8cc5-42ac-a956-4d890b023403","Type":"ContainerDied","Data":"41736078cad50daa0ef0a1d7d69b0652b8bf0cbf2e5baa3e8bc288e2f20485f6"} Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.565311 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41736078cad50daa0ef0a1d7d69b0652b8bf0cbf2e5baa3e8bc288e2f20485f6" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.565372 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d2wsq" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.569528 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-df7c-account-create-update-5f4c6" event={"ID":"71654a6c-0437-494a-9162-a1dcd1d3547f","Type":"ContainerDied","Data":"0138d3becce33181ea3a443171e78cb3d5c4476c6131353dbfebe1ae698bbe3d"} Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.569571 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0138d3becce33181ea3a443171e78cb3d5c4476c6131353dbfebe1ae698bbe3d" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.569616 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-df7c-account-create-update-5f4c6" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.575512 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4a47-account-create-update-qbfpz" event={"ID":"221a021f-7c38-4604-b717-2a93af571c05","Type":"ContainerDied","Data":"9d323c78ff596cf310893520607cf68a699555742953389d1d156e5d3337a4a9"} Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.575556 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d323c78ff596cf310893520607cf68a699555742953389d1d156e5d3337a4a9" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.575630 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4a47-account-create-update-qbfpz" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.577923 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0b3c-account-create-update-7wnbl" Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.578818 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0b3c-account-create-update-7wnbl" event={"ID":"161bc45b-0f32-488e-bda2-eb150afafcda","Type":"ContainerDied","Data":"33191f756e44007c10dd343fdc8a15bae8804ea231ab780e26b32cd0ba26036f"} Dec 11 08:57:07 crc kubenswrapper[4629]: I1211 08:57:07.578859 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33191f756e44007c10dd343fdc8a15bae8804ea231ab780e26b32cd0ba26036f" Dec 11 08:57:07 crc kubenswrapper[4629]: E1211 08:57:07.580895 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-keystone:current-podified\\\"\"" pod="openstack/keystone-db-sync-gvk6g" podUID="29518fb5-bde7-4468-860f-37b425465f70" Dec 11 08:57:18 crc kubenswrapper[4629]: I1211 08:57:18.663182 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j8tjt" event={"ID":"c469e9bc-a78a-41cc-8034-1fbec2d2452e","Type":"ContainerStarted","Data":"74e330e1f6221178db4c4901a82fdcbe28429b4d43fade140f4657a86594d26f"} Dec 11 08:57:18 crc kubenswrapper[4629]: I1211 08:57:18.683191 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-j8tjt" podStartSLOduration=2.655935633 podStartE2EDuration="37.68317104s" podCreationTimestamp="2025-12-11 08:56:41 +0000 UTC" firstStartedPulling="2025-12-11 08:56:42.629799034 +0000 UTC m=+1190.695214642" lastFinishedPulling="2025-12-11 08:57:17.657034441 +0000 UTC m=+1225.722450049" observedRunningTime="2025-12-11 08:57:18.678523693 +0000 UTC m=+1226.743939301" watchObservedRunningTime="2025-12-11 08:57:18.68317104 +0000 UTC m=+1226.748586658" Dec 11 08:57:19 crc kubenswrapper[4629]: I1211 08:57:19.676982 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gvk6g" event={"ID":"29518fb5-bde7-4468-860f-37b425465f70","Type":"ContainerStarted","Data":"6dfe4c4c6a971a1a0a70a76875dd8e775ec48a6d91dd53f9cbf82c29bc0a250e"} Dec 11 08:57:19 crc kubenswrapper[4629]: I1211 08:57:19.703758 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-gvk6g" podStartSLOduration=3.01491459 podStartE2EDuration="30.703734642s" podCreationTimestamp="2025-12-11 08:56:49 +0000 UTC" firstStartedPulling="2025-12-11 08:56:51.344648973 +0000 UTC m=+1199.410064581" lastFinishedPulling="2025-12-11 08:57:19.033469035 +0000 UTC m=+1227.098884633" observedRunningTime="2025-12-11 08:57:19.69981842 +0000 UTC m=+1227.765234028" watchObservedRunningTime="2025-12-11 08:57:19.703734642 +0000 UTC m=+1227.769150250" Dec 11 08:57:23 crc kubenswrapper[4629]: I1211 08:57:23.713159 4629 generic.go:334] "Generic (PLEG): container finished" podID="29518fb5-bde7-4468-860f-37b425465f70" containerID="6dfe4c4c6a971a1a0a70a76875dd8e775ec48a6d91dd53f9cbf82c29bc0a250e" exitCode=0 Dec 11 08:57:23 crc kubenswrapper[4629]: I1211 08:57:23.713256 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gvk6g" event={"ID":"29518fb5-bde7-4468-860f-37b425465f70","Type":"ContainerDied","Data":"6dfe4c4c6a971a1a0a70a76875dd8e775ec48a6d91dd53f9cbf82c29bc0a250e"} Dec 11 08:57:24 crc kubenswrapper[4629]: I1211 08:57:24.999105 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:57:25 crc kubenswrapper[4629]: I1211 08:57:25.197909 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29518fb5-bde7-4468-860f-37b425465f70-combined-ca-bundle\") pod \"29518fb5-bde7-4468-860f-37b425465f70\" (UID: \"29518fb5-bde7-4468-860f-37b425465f70\") " Dec 11 08:57:25 crc kubenswrapper[4629]: I1211 08:57:25.198006 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29518fb5-bde7-4468-860f-37b425465f70-config-data\") pod \"29518fb5-bde7-4468-860f-37b425465f70\" (UID: \"29518fb5-bde7-4468-860f-37b425465f70\") " Dec 11 08:57:25 crc kubenswrapper[4629]: I1211 08:57:25.198150 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pwlp\" (UniqueName: \"kubernetes.io/projected/29518fb5-bde7-4468-860f-37b425465f70-kube-api-access-8pwlp\") pod \"29518fb5-bde7-4468-860f-37b425465f70\" (UID: \"29518fb5-bde7-4468-860f-37b425465f70\") " Dec 11 08:57:25 crc kubenswrapper[4629]: I1211 08:57:25.205123 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29518fb5-bde7-4468-860f-37b425465f70-kube-api-access-8pwlp" (OuterVolumeSpecName: "kube-api-access-8pwlp") pod "29518fb5-bde7-4468-860f-37b425465f70" (UID: "29518fb5-bde7-4468-860f-37b425465f70"). InnerVolumeSpecName "kube-api-access-8pwlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:25 crc kubenswrapper[4629]: I1211 08:57:25.230521 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29518fb5-bde7-4468-860f-37b425465f70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29518fb5-bde7-4468-860f-37b425465f70" (UID: "29518fb5-bde7-4468-860f-37b425465f70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:57:25 crc kubenswrapper[4629]: I1211 08:57:25.266106 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29518fb5-bde7-4468-860f-37b425465f70-config-data" (OuterVolumeSpecName: "config-data") pod "29518fb5-bde7-4468-860f-37b425465f70" (UID: "29518fb5-bde7-4468-860f-37b425465f70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:57:25 crc kubenswrapper[4629]: I1211 08:57:25.301234 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29518fb5-bde7-4468-860f-37b425465f70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:25 crc kubenswrapper[4629]: I1211 08:57:25.301282 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29518fb5-bde7-4468-860f-37b425465f70-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:25 crc kubenswrapper[4629]: I1211 08:57:25.301294 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pwlp\" (UniqueName: \"kubernetes.io/projected/29518fb5-bde7-4468-860f-37b425465f70-kube-api-access-8pwlp\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:25 crc kubenswrapper[4629]: I1211 08:57:25.730134 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gvk6g" event={"ID":"29518fb5-bde7-4468-860f-37b425465f70","Type":"ContainerDied","Data":"c57e859bbe5feda781c486fbd775bef4aa5cfaecdf022fcba78dd0154bc5d7e8"} Dec 11 08:57:25 crc kubenswrapper[4629]: I1211 08:57:25.730187 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c57e859bbe5feda781c486fbd775bef4aa5cfaecdf022fcba78dd0154bc5d7e8" Dec 11 08:57:25 crc kubenswrapper[4629]: I1211 08:57:25.730199 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gvk6g" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.050273 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-nm9gh"] Dec 11 08:57:26 crc kubenswrapper[4629]: E1211 08:57:26.050589 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac490e45-b339-4803-b79e-075da951724b" containerName="mariadb-database-create" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.050601 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac490e45-b339-4803-b79e-075da951724b" containerName="mariadb-database-create" Dec 11 08:57:26 crc kubenswrapper[4629]: E1211 08:57:26.050614 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29518fb5-bde7-4468-860f-37b425465f70" containerName="keystone-db-sync" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.050621 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="29518fb5-bde7-4468-860f-37b425465f70" containerName="keystone-db-sync" Dec 11 08:57:26 crc kubenswrapper[4629]: E1211 08:57:26.050630 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71654a6c-0437-494a-9162-a1dcd1d3547f" containerName="mariadb-account-create-update" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.050636 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="71654a6c-0437-494a-9162-a1dcd1d3547f" containerName="mariadb-account-create-update" Dec 11 08:57:26 crc kubenswrapper[4629]: E1211 08:57:26.050652 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="221a021f-7c38-4604-b717-2a93af571c05" containerName="mariadb-account-create-update" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.050686 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="221a021f-7c38-4604-b717-2a93af571c05" containerName="mariadb-account-create-update" Dec 11 08:57:26 crc kubenswrapper[4629]: E1211 08:57:26.050712 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="161bc45b-0f32-488e-bda2-eb150afafcda" containerName="mariadb-account-create-update" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.050721 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="161bc45b-0f32-488e-bda2-eb150afafcda" containerName="mariadb-account-create-update" Dec 11 08:57:26 crc kubenswrapper[4629]: E1211 08:57:26.050732 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b387bcfc-8cc5-42ac-a956-4d890b023403" containerName="mariadb-database-create" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.050739 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="b387bcfc-8cc5-42ac-a956-4d890b023403" containerName="mariadb-database-create" Dec 11 08:57:26 crc kubenswrapper[4629]: E1211 08:57:26.050750 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ed3ed74-d821-4428-b47b-50fdc38d6bf2" containerName="mariadb-database-create" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.050756 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ed3ed74-d821-4428-b47b-50fdc38d6bf2" containerName="mariadb-database-create" Dec 11 08:57:26 crc kubenswrapper[4629]: E1211 08:57:26.050769 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="613817ce-916e-4cb2-98cf-e55bd98c27f9" containerName="ovn-config" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.050776 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="613817ce-916e-4cb2-98cf-e55bd98c27f9" containerName="ovn-config" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.051634 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="29518fb5-bde7-4468-860f-37b425465f70" containerName="keystone-db-sync" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.051656 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac490e45-b339-4803-b79e-075da951724b" containerName="mariadb-database-create" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.051670 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="161bc45b-0f32-488e-bda2-eb150afafcda" containerName="mariadb-account-create-update" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.051686 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="613817ce-916e-4cb2-98cf-e55bd98c27f9" containerName="ovn-config" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.051695 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ed3ed74-d821-4428-b47b-50fdc38d6bf2" containerName="mariadb-database-create" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.051705 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="b387bcfc-8cc5-42ac-a956-4d890b023403" containerName="mariadb-database-create" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.051715 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="221a021f-7c38-4604-b717-2a93af571c05" containerName="mariadb-account-create-update" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.051726 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="71654a6c-0437-494a-9162-a1dcd1d3547f" containerName="mariadb-account-create-update" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.052767 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.094114 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-nm9gh"] Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.100614 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-592gd"] Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.101698 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.107968 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.108317 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.108502 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.108695 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-m896j" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.109331 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.195495 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-592gd"] Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.213871 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-dns-svc\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.213928 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ng4g\" (UniqueName: \"kubernetes.io/projected/ccc0fffd-c9da-483d-a684-8193b57fce11-kube-api-access-7ng4g\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.213989 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-scripts\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.214015 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-ovsdbserver-nb\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.214048 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-config\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.214094 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvmnh\" (UniqueName: \"kubernetes.io/projected/93ed8db8-a180-4968-bc4a-cae8bf1e2962-kube-api-access-cvmnh\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.214131 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-config-data\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.214185 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-credential-keys\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.214218 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-ovsdbserver-sb\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.214271 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-combined-ca-bundle\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.214295 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-fernet-keys\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.317880 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-dns-svc\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.317927 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ng4g\" (UniqueName: \"kubernetes.io/projected/ccc0fffd-c9da-483d-a684-8193b57fce11-kube-api-access-7ng4g\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.317996 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-scripts\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.318013 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-ovsdbserver-nb\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.318035 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-config\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.318076 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvmnh\" (UniqueName: \"kubernetes.io/projected/93ed8db8-a180-4968-bc4a-cae8bf1e2962-kube-api-access-cvmnh\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.318104 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-config-data\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.318141 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-credential-keys\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.318163 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-ovsdbserver-sb\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.318205 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-combined-ca-bundle\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.318219 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-fernet-keys\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.320148 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-dns-svc\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.324138 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-ovsdbserver-sb\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.324774 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-ovsdbserver-nb\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.328046 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-config\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.333362 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-fernet-keys\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.335708 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-credential-keys\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.342942 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-combined-ca-bundle\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.475107 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-scripts\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.499817 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-config-data\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.531370 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvmnh\" (UniqueName: \"kubernetes.io/projected/93ed8db8-a180-4968-bc4a-cae8bf1e2962-kube-api-access-cvmnh\") pod \"keystone-bootstrap-592gd\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.562964 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ng4g\" (UniqueName: \"kubernetes.io/projected/ccc0fffd-c9da-483d-a684-8193b57fce11-kube-api-access-7ng4g\") pod \"dnsmasq-dns-66fbd85b65-nm9gh\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.674797 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.725525 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.755525 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-vr26n"] Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.768868 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.784830 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.785063 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-57gv4" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.785222 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.815926 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vr26n"] Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.949057 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-combined-ca-bundle\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.949145 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c6d6e77-ab1f-494b-9d78-78544b6068c4-etc-machine-id\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.949188 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn2jx\" (UniqueName: \"kubernetes.io/projected/9c6d6e77-ab1f-494b-9d78-78544b6068c4-kube-api-access-dn2jx\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.949286 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-config-data\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.949350 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-db-sync-config-data\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:26 crc kubenswrapper[4629]: I1211 08:57:26.949388 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-scripts\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.053194 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-combined-ca-bundle\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.053305 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c6d6e77-ab1f-494b-9d78-78544b6068c4-etc-machine-id\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.053358 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn2jx\" (UniqueName: \"kubernetes.io/projected/9c6d6e77-ab1f-494b-9d78-78544b6068c4-kube-api-access-dn2jx\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.053487 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-config-data\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.053591 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-db-sync-config-data\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.053649 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-scripts\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.059412 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c6d6e77-ab1f-494b-9d78-78544b6068c4-etc-machine-id\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.089890 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-config-data\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.095997 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-combined-ca-bundle\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.171915 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-scripts\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.176414 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-db-sync-config-data\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.246909 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2thgx"] Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.247924 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2thgx" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.268864 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tzz4d" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.268914 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.273713 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.316985 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn2jx\" (UniqueName: \"kubernetes.io/projected/9c6d6e77-ab1f-494b-9d78-78544b6068c4-kube-api-access-dn2jx\") pod \"cinder-db-sync-vr26n\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.331260 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2thgx"] Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.347329 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-fb8kh"] Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.423363 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.488574 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-blx9j" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.510147 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.540018 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-combined-ca-bundle\") pod \"neutron-db-sync-2thgx\" (UID: \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\") " pod="openstack/neutron-db-sync-2thgx" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.540120 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-config\") pod \"neutron-db-sync-2thgx\" (UID: \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\") " pod="openstack/neutron-db-sync-2thgx" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.540581 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9zjc\" (UniqueName: \"kubernetes.io/projected/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-kube-api-access-c9zjc\") pod \"neutron-db-sync-2thgx\" (UID: \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\") " pod="openstack/neutron-db-sync-2thgx" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.544669 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.621865 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vr26n" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.750185 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-scripts\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.750270 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-combined-ca-bundle\") pod \"neutron-db-sync-2thgx\" (UID: \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\") " pod="openstack/neutron-db-sync-2thgx" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.750289 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-config\") pod \"neutron-db-sync-2thgx\" (UID: \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\") " pod="openstack/neutron-db-sync-2thgx" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.750319 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-config-data\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.750357 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4m6j\" (UniqueName: \"kubernetes.io/projected/6b3da647-a18c-4672-9757-cbeab4f49fe0-kube-api-access-c4m6j\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.750416 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b3da647-a18c-4672-9757-cbeab4f49fe0-logs\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.750439 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-combined-ca-bundle\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.750460 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9zjc\" (UniqueName: \"kubernetes.io/projected/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-kube-api-access-c9zjc\") pod \"neutron-db-sync-2thgx\" (UID: \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\") " pod="openstack/neutron-db-sync-2thgx" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.774564 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-combined-ca-bundle\") pod \"neutron-db-sync-2thgx\" (UID: \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\") " pod="openstack/neutron-db-sync-2thgx" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.812932 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-config\") pod \"neutron-db-sync-2thgx\" (UID: \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\") " pod="openstack/neutron-db-sync-2thgx" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.837391 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9zjc\" (UniqueName: \"kubernetes.io/projected/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-kube-api-access-c9zjc\") pod \"neutron-db-sync-2thgx\" (UID: \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\") " pod="openstack/neutron-db-sync-2thgx" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.858124 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-config-data\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.858198 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4m6j\" (UniqueName: \"kubernetes.io/projected/6b3da647-a18c-4672-9757-cbeab4f49fe0-kube-api-access-c4m6j\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.858274 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b3da647-a18c-4672-9757-cbeab4f49fe0-logs\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.858308 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-combined-ca-bundle\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.858339 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-scripts\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.865495 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b3da647-a18c-4672-9757-cbeab4f49fe0-logs\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.865997 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.868351 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.869835 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-scripts\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.888447 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.889234 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-combined-ca-bundle\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.893375 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-config-data\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.907647 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.923447 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2thgx" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.963564 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.963628 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-log-httpd\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.963670 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-config-data\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.963701 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-scripts\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.963734 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-run-httpd\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.963779 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tfbq\" (UniqueName: \"kubernetes.io/projected/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-kube-api-access-7tfbq\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.963824 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.979871 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-fb8kh"] Dec 11 08:57:27 crc kubenswrapper[4629]: I1211 08:57:27.991425 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4m6j\" (UniqueName: \"kubernetes.io/projected/6b3da647-a18c-4672-9757-cbeab4f49fe0-kube-api-access-c4m6j\") pod \"placement-db-sync-fb8kh\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.050587 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-nm9gh"] Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.066340 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tfbq\" (UniqueName: \"kubernetes.io/projected/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-kube-api-access-7tfbq\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.066887 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.066997 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.067047 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-log-httpd\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.067084 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-config-data\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.067117 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-scripts\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.067154 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-run-httpd\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.067744 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-run-httpd\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.068098 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-log-httpd\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.070924 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.106829 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tfbq\" (UniqueName: \"kubernetes.io/projected/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-kube-api-access-7tfbq\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.114102 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-pdwnh"] Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.119444 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.139241 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-mtgkx"] Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.140791 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.141444 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.142031 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-k4426" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.142323 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.144824 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-scripts\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.154366 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.155882 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-config-data\") pod \"ceilometer-0\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.176236 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fb8kh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.275764 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.276280 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c198d7a3-15c2-4716-b9a2-4d008076bfdb-db-sync-config-data\") pod \"barbican-db-sync-pdwnh\" (UID: \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\") " pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.276341 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwclc\" (UniqueName: \"kubernetes.io/projected/c198d7a3-15c2-4716-b9a2-4d008076bfdb-kube-api-access-qwclc\") pod \"barbican-db-sync-pdwnh\" (UID: \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\") " pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.276415 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-dns-svc\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.276467 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-config\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.276585 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c198d7a3-15c2-4716-b9a2-4d008076bfdb-combined-ca-bundle\") pod \"barbican-db-sync-pdwnh\" (UID: \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\") " pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.276613 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drd56\" (UniqueName: \"kubernetes.io/projected/df139500-bd57-4e56-99f8-7c47b03938b9-kube-api-access-drd56\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.276676 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.290936 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pdwnh"] Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.295197 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.304090 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-mtgkx"] Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.321039 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-nm9gh"] Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.390466 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c198d7a3-15c2-4716-b9a2-4d008076bfdb-db-sync-config-data\") pod \"barbican-db-sync-pdwnh\" (UID: \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\") " pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.390542 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwclc\" (UniqueName: \"kubernetes.io/projected/c198d7a3-15c2-4716-b9a2-4d008076bfdb-kube-api-access-qwclc\") pod \"barbican-db-sync-pdwnh\" (UID: \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\") " pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.390654 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-dns-svc\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.390684 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-config\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.390748 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c198d7a3-15c2-4716-b9a2-4d008076bfdb-combined-ca-bundle\") pod \"barbican-db-sync-pdwnh\" (UID: \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\") " pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.390786 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drd56\" (UniqueName: \"kubernetes.io/projected/df139500-bd57-4e56-99f8-7c47b03938b9-kube-api-access-drd56\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.390836 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.390908 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.392793 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.393437 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-config\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.397488 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.397767 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-dns-svc\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.402791 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c198d7a3-15c2-4716-b9a2-4d008076bfdb-combined-ca-bundle\") pod \"barbican-db-sync-pdwnh\" (UID: \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\") " pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.411728 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c198d7a3-15c2-4716-b9a2-4d008076bfdb-db-sync-config-data\") pod \"barbican-db-sync-pdwnh\" (UID: \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\") " pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.438487 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwclc\" (UniqueName: \"kubernetes.io/projected/c198d7a3-15c2-4716-b9a2-4d008076bfdb-kube-api-access-qwclc\") pod \"barbican-db-sync-pdwnh\" (UID: \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\") " pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.440971 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drd56\" (UniqueName: \"kubernetes.io/projected/df139500-bd57-4e56-99f8-7c47b03938b9-kube-api-access-drd56\") pod \"dnsmasq-dns-6bf59f66bf-mtgkx\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.488192 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-592gd"] Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.513503 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.526650 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.613752 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vr26n"] Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.886308 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" event={"ID":"ccc0fffd-c9da-483d-a684-8193b57fce11","Type":"ContainerStarted","Data":"3c1514c7a6505553a60dda9002bbe522c29c31e61c732375bf557f0d54970a73"} Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.896673 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-592gd" event={"ID":"93ed8db8-a180-4968-bc4a-cae8bf1e2962","Type":"ContainerStarted","Data":"f31b0d0c957711c6b3091f8379b620c8330153d1453a576dfc9e3f8138e1beba"} Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.917768 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vr26n" event={"ID":"9c6d6e77-ab1f-494b-9d78-78544b6068c4","Type":"ContainerStarted","Data":"123c9f8e939769640846d802f236a8d41baee10da2d99b5eef4f9f3c784ffce3"} Dec 11 08:57:28 crc kubenswrapper[4629]: I1211 08:57:28.982970 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2thgx"] Dec 11 08:57:29 crc kubenswrapper[4629]: I1211 08:57:29.111762 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:57:29 crc kubenswrapper[4629]: W1211 08:57:29.154888 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b3da647_a18c_4672_9757_cbeab4f49fe0.slice/crio-0b33a9c5e831f4ff28156e87be944de66c2168f8b99e0180c9cc258c1da6f4d6 WatchSource:0}: Error finding container 0b33a9c5e831f4ff28156e87be944de66c2168f8b99e0180c9cc258c1da6f4d6: Status 404 returned error can't find the container with id 0b33a9c5e831f4ff28156e87be944de66c2168f8b99e0180c9cc258c1da6f4d6 Dec 11 08:57:29 crc kubenswrapper[4629]: I1211 08:57:29.163797 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-fb8kh"] Dec 11 08:57:29 crc kubenswrapper[4629]: I1211 08:57:29.491482 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-mtgkx"] Dec 11 08:57:29 crc kubenswrapper[4629]: I1211 08:57:29.636451 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pdwnh"] Dec 11 08:57:29 crc kubenswrapper[4629]: I1211 08:57:29.959341 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fb8kh" event={"ID":"6b3da647-a18c-4672-9757-cbeab4f49fe0","Type":"ContainerStarted","Data":"0b33a9c5e831f4ff28156e87be944de66c2168f8b99e0180c9cc258c1da6f4d6"} Dec 11 08:57:29 crc kubenswrapper[4629]: I1211 08:57:29.969151 4629 generic.go:334] "Generic (PLEG): container finished" podID="df139500-bd57-4e56-99f8-7c47b03938b9" containerID="4067b357fccbe8a92c29b45b40d4a6f746ad27d2fb59d85274870f86f60276cb" exitCode=0 Dec 11 08:57:29 crc kubenswrapper[4629]: I1211 08:57:29.969214 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" event={"ID":"df139500-bd57-4e56-99f8-7c47b03938b9","Type":"ContainerDied","Data":"4067b357fccbe8a92c29b45b40d4a6f746ad27d2fb59d85274870f86f60276cb"} Dec 11 08:57:29 crc kubenswrapper[4629]: I1211 08:57:29.969240 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" event={"ID":"df139500-bd57-4e56-99f8-7c47b03938b9","Type":"ContainerStarted","Data":"2bf515280a5461feabdf37f521e02639cf8ccc4201d74b20fc01aadd98456176"} Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:29.999140 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-592gd" event={"ID":"93ed8db8-a180-4968-bc4a-cae8bf1e2962","Type":"ContainerStarted","Data":"fabd10859759368fe09db4183b04a74db915ffa14133f4e49319ebc339cc439c"} Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.017112 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2thgx" event={"ID":"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23","Type":"ContainerStarted","Data":"851281ce78b0aadd6cf06383ef11bf1bf1d2b7d105886948b3b2bb5e499bef2a"} Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.017162 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2thgx" event={"ID":"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23","Type":"ContainerStarted","Data":"684e511da4ac48c50572b92b71b04c9f9cd0d0e7279196909810ddf75cfacab3"} Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.034183 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pdwnh" event={"ID":"c198d7a3-15c2-4716-b9a2-4d008076bfdb","Type":"ContainerStarted","Data":"763e12b36df5966d1af5a1488deb657a00e7415f864c4da114fcf556763e8b91"} Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.047548 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-592gd" podStartSLOduration=4.047527121 podStartE2EDuration="4.047527121s" podCreationTimestamp="2025-12-11 08:57:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:57:30.034359277 +0000 UTC m=+1238.099774895" watchObservedRunningTime="2025-12-11 08:57:30.047527121 +0000 UTC m=+1238.112942719" Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.064292 4629 generic.go:334] "Generic (PLEG): container finished" podID="ccc0fffd-c9da-483d-a684-8193b57fce11" containerID="8031d32cbf9cb04f5e6df4958152f56b1760c6040bab87e1142a058831a9bd04" exitCode=0 Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.064381 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" event={"ID":"ccc0fffd-c9da-483d-a684-8193b57fce11","Type":"ContainerDied","Data":"8031d32cbf9cb04f5e6df4958152f56b1760c6040bab87e1142a058831a9bd04"} Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.076726 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a73b4d6-ad71-4ea6-ab69-17d1c581b911","Type":"ContainerStarted","Data":"c8440e532dceed0c309959737566b0d8d8dadbd414c6534d0dbfe2a8e7676368"} Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.081757 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2thgx" podStartSLOduration=4.081736824 podStartE2EDuration="4.081736824s" podCreationTimestamp="2025-12-11 08:57:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:57:30.071625457 +0000 UTC m=+1238.137041065" watchObservedRunningTime="2025-12-11 08:57:30.081736824 +0000 UTC m=+1238.147152432" Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.608925 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.686896 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-ovsdbserver-nb\") pod \"ccc0fffd-c9da-483d-a684-8193b57fce11\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.686996 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-config\") pod \"ccc0fffd-c9da-483d-a684-8193b57fce11\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.687019 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ng4g\" (UniqueName: \"kubernetes.io/projected/ccc0fffd-c9da-483d-a684-8193b57fce11-kube-api-access-7ng4g\") pod \"ccc0fffd-c9da-483d-a684-8193b57fce11\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.687082 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-dns-svc\") pod \"ccc0fffd-c9da-483d-a684-8193b57fce11\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.687193 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-ovsdbserver-sb\") pod \"ccc0fffd-c9da-483d-a684-8193b57fce11\" (UID: \"ccc0fffd-c9da-483d-a684-8193b57fce11\") " Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.728119 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccc0fffd-c9da-483d-a684-8193b57fce11-kube-api-access-7ng4g" (OuterVolumeSpecName: "kube-api-access-7ng4g") pod "ccc0fffd-c9da-483d-a684-8193b57fce11" (UID: "ccc0fffd-c9da-483d-a684-8193b57fce11"). InnerVolumeSpecName "kube-api-access-7ng4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.772143 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ccc0fffd-c9da-483d-a684-8193b57fce11" (UID: "ccc0fffd-c9da-483d-a684-8193b57fce11"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.774716 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ccc0fffd-c9da-483d-a684-8193b57fce11" (UID: "ccc0fffd-c9da-483d-a684-8193b57fce11"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.780410 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-config" (OuterVolumeSpecName: "config") pod "ccc0fffd-c9da-483d-a684-8193b57fce11" (UID: "ccc0fffd-c9da-483d-a684-8193b57fce11"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.789148 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.789212 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.789239 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ng4g\" (UniqueName: \"kubernetes.io/projected/ccc0fffd-c9da-483d-a684-8193b57fce11-kube-api-access-7ng4g\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.789263 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.834161 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ccc0fffd-c9da-483d-a684-8193b57fce11" (UID: "ccc0fffd-c9da-483d-a684-8193b57fce11"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:30 crc kubenswrapper[4629]: I1211 08:57:30.891659 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccc0fffd-c9da-483d-a684-8193b57fce11-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:31 crc kubenswrapper[4629]: I1211 08:57:31.213233 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" event={"ID":"ccc0fffd-c9da-483d-a684-8193b57fce11","Type":"ContainerDied","Data":"3c1514c7a6505553a60dda9002bbe522c29c31e61c732375bf557f0d54970a73"} Dec 11 08:57:31 crc kubenswrapper[4629]: I1211 08:57:31.213299 4629 scope.go:117] "RemoveContainer" containerID="8031d32cbf9cb04f5e6df4958152f56b1760c6040bab87e1142a058831a9bd04" Dec 11 08:57:31 crc kubenswrapper[4629]: I1211 08:57:31.213467 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fbd85b65-nm9gh" Dec 11 08:57:31 crc kubenswrapper[4629]: I1211 08:57:31.242713 4629 generic.go:334] "Generic (PLEG): container finished" podID="c469e9bc-a78a-41cc-8034-1fbec2d2452e" containerID="74e330e1f6221178db4c4901a82fdcbe28429b4d43fade140f4657a86594d26f" exitCode=0 Dec 11 08:57:31 crc kubenswrapper[4629]: I1211 08:57:31.242835 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j8tjt" event={"ID":"c469e9bc-a78a-41cc-8034-1fbec2d2452e","Type":"ContainerDied","Data":"74e330e1f6221178db4c4901a82fdcbe28429b4d43fade140f4657a86594d26f"} Dec 11 08:57:31 crc kubenswrapper[4629]: I1211 08:57:31.315177 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" event={"ID":"df139500-bd57-4e56-99f8-7c47b03938b9","Type":"ContainerStarted","Data":"488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e"} Dec 11 08:57:31 crc kubenswrapper[4629]: I1211 08:57:31.315897 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:31 crc kubenswrapper[4629]: I1211 08:57:31.416996 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-nm9gh"] Dec 11 08:57:31 crc kubenswrapper[4629]: I1211 08:57:31.422508 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-nm9gh"] Dec 11 08:57:31 crc kubenswrapper[4629]: I1211 08:57:31.984599 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" podStartSLOduration=4.98456414 podStartE2EDuration="4.98456414s" podCreationTimestamp="2025-12-11 08:57:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:57:31.470253996 +0000 UTC m=+1239.535669604" watchObservedRunningTime="2025-12-11 08:57:31.98456414 +0000 UTC m=+1240.049979738" Dec 11 08:57:31 crc kubenswrapper[4629]: I1211 08:57:31.994340 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:57:32 crc kubenswrapper[4629]: I1211 08:57:32.297566 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccc0fffd-c9da-483d-a684-8193b57fce11" path="/var/lib/kubelet/pods/ccc0fffd-c9da-483d-a684-8193b57fce11/volumes" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.203029 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j8tjt" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.388165 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j8tjt" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.388365 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j8tjt" event={"ID":"c469e9bc-a78a-41cc-8034-1fbec2d2452e","Type":"ContainerDied","Data":"01cb5f27edebfe7ca2bf195860c56f52149736b44807d0f0c8d186970788efe9"} Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.388389 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01cb5f27edebfe7ca2bf195860c56f52149736b44807d0f0c8d186970788efe9" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.400732 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tc7n\" (UniqueName: \"kubernetes.io/projected/c469e9bc-a78a-41cc-8034-1fbec2d2452e-kube-api-access-8tc7n\") pod \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.403641 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-combined-ca-bundle\") pod \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.403789 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-db-sync-config-data\") pod \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.403832 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-config-data\") pod \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\" (UID: \"c469e9bc-a78a-41cc-8034-1fbec2d2452e\") " Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.419167 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c469e9bc-a78a-41cc-8034-1fbec2d2452e" (UID: "c469e9bc-a78a-41cc-8034-1fbec2d2452e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.463512 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c469e9bc-a78a-41cc-8034-1fbec2d2452e-kube-api-access-8tc7n" (OuterVolumeSpecName: "kube-api-access-8tc7n") pod "c469e9bc-a78a-41cc-8034-1fbec2d2452e" (UID: "c469e9bc-a78a-41cc-8034-1fbec2d2452e"). InnerVolumeSpecName "kube-api-access-8tc7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.482751 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c469e9bc-a78a-41cc-8034-1fbec2d2452e" (UID: "c469e9bc-a78a-41cc-8034-1fbec2d2452e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.505983 4629 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.506191 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tc7n\" (UniqueName: \"kubernetes.io/projected/c469e9bc-a78a-41cc-8034-1fbec2d2452e-kube-api-access-8tc7n\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.506277 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.547780 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-config-data" (OuterVolumeSpecName: "config-data") pod "c469e9bc-a78a-41cc-8034-1fbec2d2452e" (UID: "c469e9bc-a78a-41cc-8034-1fbec2d2452e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.614284 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c469e9bc-a78a-41cc-8034-1fbec2d2452e-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.850192 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-mtgkx"] Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.928409 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp"] Dec 11 08:57:33 crc kubenswrapper[4629]: E1211 08:57:33.928796 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc0fffd-c9da-483d-a684-8193b57fce11" containerName="init" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.928819 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc0fffd-c9da-483d-a684-8193b57fce11" containerName="init" Dec 11 08:57:33 crc kubenswrapper[4629]: E1211 08:57:33.928870 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c469e9bc-a78a-41cc-8034-1fbec2d2452e" containerName="glance-db-sync" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.928877 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c469e9bc-a78a-41cc-8034-1fbec2d2452e" containerName="glance-db-sync" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.932836 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccc0fffd-c9da-483d-a684-8193b57fce11" containerName="init" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.932894 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="c469e9bc-a78a-41cc-8034-1fbec2d2452e" containerName="glance-db-sync" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.933940 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:33 crc kubenswrapper[4629]: I1211 08:57:33.960688 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp"] Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.025624 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.025718 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-config\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.025769 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.026009 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.026071 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzt9t\" (UniqueName: \"kubernetes.io/projected/297d67cc-431f-45ae-8e51-6df1148414d2-kube-api-access-lzt9t\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.127986 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.128045 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-config\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.128062 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.128115 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.128135 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzt9t\" (UniqueName: \"kubernetes.io/projected/297d67cc-431f-45ae-8e51-6df1148414d2-kube-api-access-lzt9t\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.129237 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.129328 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-config\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.129684 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.130050 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.181778 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzt9t\" (UniqueName: \"kubernetes.io/projected/297d67cc-431f-45ae-8e51-6df1148414d2-kube-api-access-lzt9t\") pod \"dnsmasq-dns-5b6dbdb6f5-vmkcp\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.266503 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.402465 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" podUID="df139500-bd57-4e56-99f8-7c47b03938b9" containerName="dnsmasq-dns" containerID="cri-o://488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e" gracePeriod=10 Dec 11 08:57:34 crc kubenswrapper[4629]: I1211 08:57:34.995335 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp"] Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.440065 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.504027 4629 generic.go:334] "Generic (PLEG): container finished" podID="df139500-bd57-4e56-99f8-7c47b03938b9" containerID="488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e" exitCode=0 Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.504126 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" event={"ID":"df139500-bd57-4e56-99f8-7c47b03938b9","Type":"ContainerDied","Data":"488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e"} Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.504154 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" event={"ID":"df139500-bd57-4e56-99f8-7c47b03938b9","Type":"ContainerDied","Data":"2bf515280a5461feabdf37f521e02639cf8ccc4201d74b20fc01aadd98456176"} Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.504170 4629 scope.go:117] "RemoveContainer" containerID="488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.504296 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf59f66bf-mtgkx" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.519308 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" event={"ID":"297d67cc-431f-45ae-8e51-6df1148414d2","Type":"ContainerStarted","Data":"ec64bc9fe42e683608f0d3dcd0de6271ac7addc05defcccc6e245ec725dbadff"} Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.576023 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-dns-svc\") pod \"df139500-bd57-4e56-99f8-7c47b03938b9\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.576096 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drd56\" (UniqueName: \"kubernetes.io/projected/df139500-bd57-4e56-99f8-7c47b03938b9-kube-api-access-drd56\") pod \"df139500-bd57-4e56-99f8-7c47b03938b9\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.576142 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-ovsdbserver-nb\") pod \"df139500-bd57-4e56-99f8-7c47b03938b9\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.576194 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-config\") pod \"df139500-bd57-4e56-99f8-7c47b03938b9\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.576554 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-ovsdbserver-sb\") pod \"df139500-bd57-4e56-99f8-7c47b03938b9\" (UID: \"df139500-bd57-4e56-99f8-7c47b03938b9\") " Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.599334 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df139500-bd57-4e56-99f8-7c47b03938b9-kube-api-access-drd56" (OuterVolumeSpecName: "kube-api-access-drd56") pod "df139500-bd57-4e56-99f8-7c47b03938b9" (UID: "df139500-bd57-4e56-99f8-7c47b03938b9"). InnerVolumeSpecName "kube-api-access-drd56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.664862 4629 scope.go:117] "RemoveContainer" containerID="4067b357fccbe8a92c29b45b40d4a6f746ad27d2fb59d85274870f86f60276cb" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.670659 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-config" (OuterVolumeSpecName: "config") pod "df139500-bd57-4e56-99f8-7c47b03938b9" (UID: "df139500-bd57-4e56-99f8-7c47b03938b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.679921 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drd56\" (UniqueName: \"kubernetes.io/projected/df139500-bd57-4e56-99f8-7c47b03938b9-kube-api-access-drd56\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.679958 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.692210 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "df139500-bd57-4e56-99f8-7c47b03938b9" (UID: "df139500-bd57-4e56-99f8-7c47b03938b9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.696139 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "df139500-bd57-4e56-99f8-7c47b03938b9" (UID: "df139500-bd57-4e56-99f8-7c47b03938b9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.706458 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "df139500-bd57-4e56-99f8-7c47b03938b9" (UID: "df139500-bd57-4e56-99f8-7c47b03938b9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.728549 4629 scope.go:117] "RemoveContainer" containerID="488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e" Dec 11 08:57:35 crc kubenswrapper[4629]: E1211 08:57:35.730739 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e\": container with ID starting with 488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e not found: ID does not exist" containerID="488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.730827 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e"} err="failed to get container status \"488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e\": rpc error: code = NotFound desc = could not find container \"488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e\": container with ID starting with 488e63cb5eaa5f3ca62ff59b4e53a897e0c79ab48f6fc7e6b41f3473555eee7e not found: ID does not exist" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.730912 4629 scope.go:117] "RemoveContainer" containerID="4067b357fccbe8a92c29b45b40d4a6f746ad27d2fb59d85274870f86f60276cb" Dec 11 08:57:35 crc kubenswrapper[4629]: E1211 08:57:35.731414 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4067b357fccbe8a92c29b45b40d4a6f746ad27d2fb59d85274870f86f60276cb\": container with ID starting with 4067b357fccbe8a92c29b45b40d4a6f746ad27d2fb59d85274870f86f60276cb not found: ID does not exist" containerID="4067b357fccbe8a92c29b45b40d4a6f746ad27d2fb59d85274870f86f60276cb" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.731461 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4067b357fccbe8a92c29b45b40d4a6f746ad27d2fb59d85274870f86f60276cb"} err="failed to get container status \"4067b357fccbe8a92c29b45b40d4a6f746ad27d2fb59d85274870f86f60276cb\": rpc error: code = NotFound desc = could not find container \"4067b357fccbe8a92c29b45b40d4a6f746ad27d2fb59d85274870f86f60276cb\": container with ID starting with 4067b357fccbe8a92c29b45b40d4a6f746ad27d2fb59d85274870f86f60276cb not found: ID does not exist" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.781823 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.781875 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.781887 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df139500-bd57-4e56-99f8-7c47b03938b9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.860104 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-mtgkx"] Dec 11 08:57:35 crc kubenswrapper[4629]: I1211 08:57:35.875905 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-mtgkx"] Dec 11 08:57:36 crc kubenswrapper[4629]: I1211 08:57:36.225436 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df139500-bd57-4e56-99f8-7c47b03938b9" path="/var/lib/kubelet/pods/df139500-bd57-4e56-99f8-7c47b03938b9/volumes" Dec 11 08:57:36 crc kubenswrapper[4629]: I1211 08:57:36.636375 4629 generic.go:334] "Generic (PLEG): container finished" podID="297d67cc-431f-45ae-8e51-6df1148414d2" containerID="c095b56340781885392fda07d1d427c63cfe63b8f15e1a258829087e0b6dac89" exitCode=0 Dec 11 08:57:36 crc kubenswrapper[4629]: I1211 08:57:36.636442 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" event={"ID":"297d67cc-431f-45ae-8e51-6df1148414d2","Type":"ContainerDied","Data":"c095b56340781885392fda07d1d427c63cfe63b8f15e1a258829087e0b6dac89"} Dec 11 08:57:37 crc kubenswrapper[4629]: I1211 08:57:37.650600 4629 generic.go:334] "Generic (PLEG): container finished" podID="93ed8db8-a180-4968-bc4a-cae8bf1e2962" containerID="fabd10859759368fe09db4183b04a74db915ffa14133f4e49319ebc339cc439c" exitCode=0 Dec 11 08:57:37 crc kubenswrapper[4629]: I1211 08:57:37.650963 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-592gd" event={"ID":"93ed8db8-a180-4968-bc4a-cae8bf1e2962","Type":"ContainerDied","Data":"fabd10859759368fe09db4183b04a74db915ffa14133f4e49319ebc339cc439c"} Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.076596 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.179522 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-scripts\") pod \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.179623 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-config-data\") pod \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.179722 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-credential-keys\") pod \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.179758 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-fernet-keys\") pod \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.179830 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-combined-ca-bundle\") pod \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.179866 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvmnh\" (UniqueName: \"kubernetes.io/projected/93ed8db8-a180-4968-bc4a-cae8bf1e2962-kube-api-access-cvmnh\") pod \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\" (UID: \"93ed8db8-a180-4968-bc4a-cae8bf1e2962\") " Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.193036 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93ed8db8-a180-4968-bc4a-cae8bf1e2962-kube-api-access-cvmnh" (OuterVolumeSpecName: "kube-api-access-cvmnh") pod "93ed8db8-a180-4968-bc4a-cae8bf1e2962" (UID: "93ed8db8-a180-4968-bc4a-cae8bf1e2962"). InnerVolumeSpecName "kube-api-access-cvmnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.199206 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-scripts" (OuterVolumeSpecName: "scripts") pod "93ed8db8-a180-4968-bc4a-cae8bf1e2962" (UID: "93ed8db8-a180-4968-bc4a-cae8bf1e2962"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.199312 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "93ed8db8-a180-4968-bc4a-cae8bf1e2962" (UID: "93ed8db8-a180-4968-bc4a-cae8bf1e2962"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.208063 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "93ed8db8-a180-4968-bc4a-cae8bf1e2962" (UID: "93ed8db8-a180-4968-bc4a-cae8bf1e2962"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.220143 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93ed8db8-a180-4968-bc4a-cae8bf1e2962" (UID: "93ed8db8-a180-4968-bc4a-cae8bf1e2962"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.230095 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-config-data" (OuterVolumeSpecName: "config-data") pod "93ed8db8-a180-4968-bc4a-cae8bf1e2962" (UID: "93ed8db8-a180-4968-bc4a-cae8bf1e2962"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.281413 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.281449 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.281462 4629 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.281476 4629 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.281488 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ed8db8-a180-4968-bc4a-cae8bf1e2962-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.281501 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvmnh\" (UniqueName: \"kubernetes.io/projected/93ed8db8-a180-4968-bc4a-cae8bf1e2962-kube-api-access-cvmnh\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:40 crc kubenswrapper[4629]: E1211 08:57:40.396570 4629 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ed8db8_a180_4968_bc4a_cae8bf1e2962.slice/crio-f31b0d0c957711c6b3091f8379b620c8330153d1453a576dfc9e3f8138e1beba\": RecentStats: unable to find data in memory cache]" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.692977 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-592gd" event={"ID":"93ed8db8-a180-4968-bc4a-cae8bf1e2962","Type":"ContainerDied","Data":"f31b0d0c957711c6b3091f8379b620c8330153d1453a576dfc9e3f8138e1beba"} Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.693612 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f31b0d0c957711c6b3091f8379b620c8330153d1453a576dfc9e3f8138e1beba" Dec 11 08:57:40 crc kubenswrapper[4629]: I1211 08:57:40.693701 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-592gd" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.215163 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-592gd"] Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.243072 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-592gd"] Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.305577 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-4sgng"] Dec 11 08:57:41 crc kubenswrapper[4629]: E1211 08:57:41.306028 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df139500-bd57-4e56-99f8-7c47b03938b9" containerName="dnsmasq-dns" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.306051 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="df139500-bd57-4e56-99f8-7c47b03938b9" containerName="dnsmasq-dns" Dec 11 08:57:41 crc kubenswrapper[4629]: E1211 08:57:41.306098 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df139500-bd57-4e56-99f8-7c47b03938b9" containerName="init" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.306107 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="df139500-bd57-4e56-99f8-7c47b03938b9" containerName="init" Dec 11 08:57:41 crc kubenswrapper[4629]: E1211 08:57:41.306122 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93ed8db8-a180-4968-bc4a-cae8bf1e2962" containerName="keystone-bootstrap" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.306132 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="93ed8db8-a180-4968-bc4a-cae8bf1e2962" containerName="keystone-bootstrap" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.306326 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="93ed8db8-a180-4968-bc4a-cae8bf1e2962" containerName="keystone-bootstrap" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.306345 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="df139500-bd57-4e56-99f8-7c47b03938b9" containerName="dnsmasq-dns" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.307020 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.310383 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.312584 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.314295 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-m896j" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.314489 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.316305 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.330957 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4sgng"] Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.579946 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-combined-ca-bundle\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.580017 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fjxb\" (UniqueName: \"kubernetes.io/projected/1cd464f4-641f-44e9-aebd-8e4766e6185a-kube-api-access-4fjxb\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.580071 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-config-data\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.580097 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-credential-keys\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.580120 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-fernet-keys\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.580171 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-scripts\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.681606 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-combined-ca-bundle\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.681674 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fjxb\" (UniqueName: \"kubernetes.io/projected/1cd464f4-641f-44e9-aebd-8e4766e6185a-kube-api-access-4fjxb\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.681731 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-config-data\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.681760 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-credential-keys\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.681782 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-fernet-keys\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.681831 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-scripts\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.694695 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-scripts\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.694840 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-credential-keys\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.697604 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-fernet-keys\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.700540 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-combined-ca-bundle\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.707870 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-config-data\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.754600 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fjxb\" (UniqueName: \"kubernetes.io/projected/1cd464f4-641f-44e9-aebd-8e4766e6185a-kube-api-access-4fjxb\") pod \"keystone-bootstrap-4sgng\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:41 crc kubenswrapper[4629]: I1211 08:57:41.924516 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:57:42 crc kubenswrapper[4629]: I1211 08:57:42.210030 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93ed8db8-a180-4968-bc4a-cae8bf1e2962" path="/var/lib/kubelet/pods/93ed8db8-a180-4968-bc4a-cae8bf1e2962/volumes" Dec 11 08:57:46 crc kubenswrapper[4629]: E1211 08:57:46.494924 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 11 08:57:46 crc kubenswrapper[4629]: E1211 08:57:46.495736 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8h5cfhb5h55h58ch75h644h546h5dhc9hf6h548h647hdh579h5d4h5fbh64bh55h577h95hb9h568h5f9h5cfh674h5f5h65fh5c8hcdhdchd9q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7tfbq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(1a73b4d6-ad71-4ea6-ab69-17d1c581b911): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:57:54 crc kubenswrapper[4629]: E1211 08:57:54.931494 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 11 08:57:54 crc kubenswrapper[4629]: E1211 08:57:54.932457 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qwclc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-pdwnh_openstack(c198d7a3-15c2-4716-b9a2-4d008076bfdb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:57:54 crc kubenswrapper[4629]: E1211 08:57:54.933671 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-pdwnh" podUID="c198d7a3-15c2-4716-b9a2-4d008076bfdb" Dec 11 08:57:55 crc kubenswrapper[4629]: E1211 08:57:55.848744 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-pdwnh" podUID="c198d7a3-15c2-4716-b9a2-4d008076bfdb" Dec 11 08:57:56 crc kubenswrapper[4629]: E1211 08:57:56.125398 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 11 08:57:56 crc kubenswrapper[4629]: E1211 08:57:56.125781 4629 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dn2jx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-vr26n_openstack(9c6d6e77-ab1f-494b-9d78-78544b6068c4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:57:56 crc kubenswrapper[4629]: E1211 08:57:56.127176 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-vr26n" podUID="9c6d6e77-ab1f-494b-9d78-78544b6068c4" Dec 11 08:57:56 crc kubenswrapper[4629]: I1211 08:57:56.616893 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4sgng"] Dec 11 08:57:56 crc kubenswrapper[4629]: W1211 08:57:56.625918 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1cd464f4_641f_44e9_aebd_8e4766e6185a.slice/crio-914135b36d387e2157f6ac70542ee405367d1e192e51be0f6391451ad35a3aae WatchSource:0}: Error finding container 914135b36d387e2157f6ac70542ee405367d1e192e51be0f6391451ad35a3aae: Status 404 returned error can't find the container with id 914135b36d387e2157f6ac70542ee405367d1e192e51be0f6391451ad35a3aae Dec 11 08:57:56 crc kubenswrapper[4629]: I1211 08:57:56.633621 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 11 08:57:56 crc kubenswrapper[4629]: I1211 08:57:56.856581 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" event={"ID":"297d67cc-431f-45ae-8e51-6df1148414d2","Type":"ContainerStarted","Data":"f872cc4ce3e0b88b5c46ab7cf729acec7266d8fe8a25f1c94ddfa5d4a60f3b4e"} Dec 11 08:57:56 crc kubenswrapper[4629]: I1211 08:57:56.857000 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:57:56 crc kubenswrapper[4629]: I1211 08:57:56.857980 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4sgng" event={"ID":"1cd464f4-641f-44e9-aebd-8e4766e6185a","Type":"ContainerStarted","Data":"986b110b28a51ad0ab15b39f0509c6e1fee8d6c32c251fb63fee8c4d441b78f9"} Dec 11 08:57:56 crc kubenswrapper[4629]: I1211 08:57:56.858002 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4sgng" event={"ID":"1cd464f4-641f-44e9-aebd-8e4766e6185a","Type":"ContainerStarted","Data":"914135b36d387e2157f6ac70542ee405367d1e192e51be0f6391451ad35a3aae"} Dec 11 08:57:56 crc kubenswrapper[4629]: I1211 08:57:56.859311 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a73b4d6-ad71-4ea6-ab69-17d1c581b911","Type":"ContainerStarted","Data":"44d6f0deeece43bb1b79e87e7a318cc1e99e9f992cb9d9bf865246d992cc0b60"} Dec 11 08:57:56 crc kubenswrapper[4629]: I1211 08:57:56.861659 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fb8kh" event={"ID":"6b3da647-a18c-4672-9757-cbeab4f49fe0","Type":"ContainerStarted","Data":"477b0a96d0245e344b8d39c2a0d564c7dc703a5c8a8fcc462dce9d2e2cc4ae49"} Dec 11 08:57:56 crc kubenswrapper[4629]: E1211 08:57:56.862925 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-vr26n" podUID="9c6d6e77-ab1f-494b-9d78-78544b6068c4" Dec 11 08:57:56 crc kubenswrapper[4629]: I1211 08:57:56.892151 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" podStartSLOduration=23.892129313 podStartE2EDuration="23.892129313s" podCreationTimestamp="2025-12-11 08:57:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:57:56.883218404 +0000 UTC m=+1264.948634012" watchObservedRunningTime="2025-12-11 08:57:56.892129313 +0000 UTC m=+1264.957544921" Dec 11 08:57:56 crc kubenswrapper[4629]: I1211 08:57:56.934389 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-fb8kh" podStartSLOduration=3.00514702 podStartE2EDuration="29.934364389s" podCreationTimestamp="2025-12-11 08:57:27 +0000 UTC" firstStartedPulling="2025-12-11 08:57:29.175485419 +0000 UTC m=+1237.240901027" lastFinishedPulling="2025-12-11 08:57:56.104702788 +0000 UTC m=+1264.170118396" observedRunningTime="2025-12-11 08:57:56.917108047 +0000 UTC m=+1264.982523655" watchObservedRunningTime="2025-12-11 08:57:56.934364389 +0000 UTC m=+1264.999779997" Dec 11 08:57:56 crc kubenswrapper[4629]: I1211 08:57:56.972837 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-4sgng" podStartSLOduration=15.972813816 podStartE2EDuration="15.972813816s" podCreationTimestamp="2025-12-11 08:57:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:57:56.968436738 +0000 UTC m=+1265.033852346" watchObservedRunningTime="2025-12-11 08:57:56.972813816 +0000 UTC m=+1265.038229424" Dec 11 08:57:59 crc kubenswrapper[4629]: I1211 08:57:59.897021 4629 generic.go:334] "Generic (PLEG): container finished" podID="6b3da647-a18c-4672-9757-cbeab4f49fe0" containerID="477b0a96d0245e344b8d39c2a0d564c7dc703a5c8a8fcc462dce9d2e2cc4ae49" exitCode=0 Dec 11 08:57:59 crc kubenswrapper[4629]: I1211 08:57:59.897240 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fb8kh" event={"ID":"6b3da647-a18c-4672-9757-cbeab4f49fe0","Type":"ContainerDied","Data":"477b0a96d0245e344b8d39c2a0d564c7dc703a5c8a8fcc462dce9d2e2cc4ae49"} Dec 11 08:58:00 crc kubenswrapper[4629]: I1211 08:58:00.937337 4629 generic.go:334] "Generic (PLEG): container finished" podID="1cd464f4-641f-44e9-aebd-8e4766e6185a" containerID="986b110b28a51ad0ab15b39f0509c6e1fee8d6c32c251fb63fee8c4d441b78f9" exitCode=0 Dec 11 08:58:00 crc kubenswrapper[4629]: I1211 08:58:00.937546 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4sgng" event={"ID":"1cd464f4-641f-44e9-aebd-8e4766e6185a","Type":"ContainerDied","Data":"986b110b28a51ad0ab15b39f0509c6e1fee8d6c32c251fb63fee8c4d441b78f9"} Dec 11 08:58:01 crc kubenswrapper[4629]: I1211 08:58:01.945720 4629 generic.go:334] "Generic (PLEG): container finished" podID="ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23" containerID="851281ce78b0aadd6cf06383ef11bf1bf1d2b7d105886948b3b2bb5e499bef2a" exitCode=0 Dec 11 08:58:01 crc kubenswrapper[4629]: I1211 08:58:01.946379 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2thgx" event={"ID":"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23","Type":"ContainerDied","Data":"851281ce78b0aadd6cf06383ef11bf1bf1d2b7d105886948b3b2bb5e499bef2a"} Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.589987 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fb8kh" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.599545 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.717660 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4m6j\" (UniqueName: \"kubernetes.io/projected/6b3da647-a18c-4672-9757-cbeab4f49fe0-kube-api-access-c4m6j\") pod \"6b3da647-a18c-4672-9757-cbeab4f49fe0\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.717730 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-config-data\") pod \"6b3da647-a18c-4672-9757-cbeab4f49fe0\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.717760 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-config-data\") pod \"1cd464f4-641f-44e9-aebd-8e4766e6185a\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.717785 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-combined-ca-bundle\") pod \"6b3da647-a18c-4672-9757-cbeab4f49fe0\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.717878 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-combined-ca-bundle\") pod \"1cd464f4-641f-44e9-aebd-8e4766e6185a\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.717911 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-scripts\") pod \"6b3da647-a18c-4672-9757-cbeab4f49fe0\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.717948 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fjxb\" (UniqueName: \"kubernetes.io/projected/1cd464f4-641f-44e9-aebd-8e4766e6185a-kube-api-access-4fjxb\") pod \"1cd464f4-641f-44e9-aebd-8e4766e6185a\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.717973 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-credential-keys\") pod \"1cd464f4-641f-44e9-aebd-8e4766e6185a\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.718007 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-fernet-keys\") pod \"1cd464f4-641f-44e9-aebd-8e4766e6185a\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.718094 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b3da647-a18c-4672-9757-cbeab4f49fe0-logs\") pod \"6b3da647-a18c-4672-9757-cbeab4f49fe0\" (UID: \"6b3da647-a18c-4672-9757-cbeab4f49fe0\") " Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.718160 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-scripts\") pod \"1cd464f4-641f-44e9-aebd-8e4766e6185a\" (UID: \"1cd464f4-641f-44e9-aebd-8e4766e6185a\") " Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.718970 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b3da647-a18c-4672-9757-cbeab4f49fe0-logs" (OuterVolumeSpecName: "logs") pod "6b3da647-a18c-4672-9757-cbeab4f49fe0" (UID: "6b3da647-a18c-4672-9757-cbeab4f49fe0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.723060 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b3da647-a18c-4672-9757-cbeab4f49fe0-kube-api-access-c4m6j" (OuterVolumeSpecName: "kube-api-access-c4m6j") pod "6b3da647-a18c-4672-9757-cbeab4f49fe0" (UID: "6b3da647-a18c-4672-9757-cbeab4f49fe0"). InnerVolumeSpecName "kube-api-access-c4m6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.730424 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1cd464f4-641f-44e9-aebd-8e4766e6185a" (UID: "1cd464f4-641f-44e9-aebd-8e4766e6185a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.730732 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-scripts" (OuterVolumeSpecName: "scripts") pod "1cd464f4-641f-44e9-aebd-8e4766e6185a" (UID: "1cd464f4-641f-44e9-aebd-8e4766e6185a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.732692 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-scripts" (OuterVolumeSpecName: "scripts") pod "6b3da647-a18c-4672-9757-cbeab4f49fe0" (UID: "6b3da647-a18c-4672-9757-cbeab4f49fe0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.732752 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1cd464f4-641f-44e9-aebd-8e4766e6185a" (UID: "1cd464f4-641f-44e9-aebd-8e4766e6185a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.734909 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cd464f4-641f-44e9-aebd-8e4766e6185a-kube-api-access-4fjxb" (OuterVolumeSpecName: "kube-api-access-4fjxb") pod "1cd464f4-641f-44e9-aebd-8e4766e6185a" (UID: "1cd464f4-641f-44e9-aebd-8e4766e6185a"). InnerVolumeSpecName "kube-api-access-4fjxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.745773 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-config-data" (OuterVolumeSpecName: "config-data") pod "6b3da647-a18c-4672-9757-cbeab4f49fe0" (UID: "6b3da647-a18c-4672-9757-cbeab4f49fe0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.747313 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1cd464f4-641f-44e9-aebd-8e4766e6185a" (UID: "1cd464f4-641f-44e9-aebd-8e4766e6185a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.748207 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b3da647-a18c-4672-9757-cbeab4f49fe0" (UID: "6b3da647-a18c-4672-9757-cbeab4f49fe0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.754915 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-config-data" (OuterVolumeSpecName: "config-data") pod "1cd464f4-641f-44e9-aebd-8e4766e6185a" (UID: "1cd464f4-641f-44e9-aebd-8e4766e6185a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.820187 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.820436 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4m6j\" (UniqueName: \"kubernetes.io/projected/6b3da647-a18c-4672-9757-cbeab4f49fe0-kube-api-access-c4m6j\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.820449 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.820461 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.820469 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.820504 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.820513 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3da647-a18c-4672-9757-cbeab4f49fe0-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.820522 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fjxb\" (UniqueName: \"kubernetes.io/projected/1cd464f4-641f-44e9-aebd-8e4766e6185a-kube-api-access-4fjxb\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.820531 4629 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.820538 4629 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1cd464f4-641f-44e9-aebd-8e4766e6185a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.820550 4629 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b3da647-a18c-4672-9757-cbeab4f49fe0-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.963466 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fb8kh" event={"ID":"6b3da647-a18c-4672-9757-cbeab4f49fe0","Type":"ContainerDied","Data":"0b33a9c5e831f4ff28156e87be944de66c2168f8b99e0180c9cc258c1da6f4d6"} Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.963520 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b33a9c5e831f4ff28156e87be944de66c2168f8b99e0180c9cc258c1da6f4d6" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.963529 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fb8kh" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.965458 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a73b4d6-ad71-4ea6-ab69-17d1c581b911","Type":"ContainerStarted","Data":"a03460dd9d0276ab2a93e7fc6d9666994f9e74e8276b474ef8cdc3c6026bfb01"} Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.971097 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4sgng" Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.973213 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4sgng" event={"ID":"1cd464f4-641f-44e9-aebd-8e4766e6185a","Type":"ContainerDied","Data":"914135b36d387e2157f6ac70542ee405367d1e192e51be0f6391451ad35a3aae"} Dec 11 08:58:02 crc kubenswrapper[4629]: I1211 08:58:02.973264 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="914135b36d387e2157f6ac70542ee405367d1e192e51be0f6391451ad35a3aae" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.066894 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5fccc9499c-f4wbp"] Dec 11 08:58:03 crc kubenswrapper[4629]: E1211 08:58:03.067353 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b3da647-a18c-4672-9757-cbeab4f49fe0" containerName="placement-db-sync" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.067376 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b3da647-a18c-4672-9757-cbeab4f49fe0" containerName="placement-db-sync" Dec 11 08:58:03 crc kubenswrapper[4629]: E1211 08:58:03.067416 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cd464f4-641f-44e9-aebd-8e4766e6185a" containerName="keystone-bootstrap" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.067426 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cd464f4-641f-44e9-aebd-8e4766e6185a" containerName="keystone-bootstrap" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.067620 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cd464f4-641f-44e9-aebd-8e4766e6185a" containerName="keystone-bootstrap" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.067646 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b3da647-a18c-4672-9757-cbeab4f49fe0" containerName="placement-db-sync" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.068748 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.074503 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.074776 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.075003 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.075260 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.075406 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.075636 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-m896j" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.081695 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5fccc9499c-f4wbp"] Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.230536 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-public-tls-certs\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.230637 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-internal-tls-certs\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.230684 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-fernet-keys\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.230715 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-scripts\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.230739 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-config-data\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.230781 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-credential-keys\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.230813 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-combined-ca-bundle\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:03 crc kubenswrapper[4629]: I1211 08:58:03.230913 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lb9g\" (UniqueName: \"kubernetes.io/projected/2998e2d9-9cba-4c82-ac41-9080e89798a3-kube-api-access-9lb9g\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.332145 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lb9g\" (UniqueName: \"kubernetes.io/projected/2998e2d9-9cba-4c82-ac41-9080e89798a3-kube-api-access-9lb9g\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.332247 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-public-tls-certs\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.332284 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-internal-tls-certs\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.332333 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-fernet-keys\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.332368 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-scripts\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.332396 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-config-data\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.332440 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-credential-keys\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.332466 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-combined-ca-bundle\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.340808 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-public-tls-certs\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.345797 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-fernet-keys\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.348926 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-credential-keys\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.349031 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-combined-ca-bundle\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.349170 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-scripts\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.349412 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-internal-tls-certs\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.351137 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2998e2d9-9cba-4c82-ac41-9080e89798a3-config-data\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.356522 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lb9g\" (UniqueName: \"kubernetes.io/projected/2998e2d9-9cba-4c82-ac41-9080e89798a3-kube-api-access-9lb9g\") pod \"keystone-5fccc9499c-f4wbp\" (UID: \"2998e2d9-9cba-4c82-ac41-9080e89798a3\") " pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.380051 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2thgx" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.439449 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.537336 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-config\") pod \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\" (UID: \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\") " Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.537809 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-combined-ca-bundle\") pod \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\" (UID: \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\") " Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.537927 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9zjc\" (UniqueName: \"kubernetes.io/projected/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-kube-api-access-c9zjc\") pod \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\" (UID: \"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23\") " Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.552156 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-kube-api-access-c9zjc" (OuterVolumeSpecName: "kube-api-access-c9zjc") pod "ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23" (UID: "ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23"). InnerVolumeSpecName "kube-api-access-c9zjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.574105 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23" (UID: "ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.583013 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-config" (OuterVolumeSpecName: "config") pod "ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23" (UID: "ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.641039 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.641084 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.641097 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9zjc\" (UniqueName: \"kubernetes.io/projected/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23-kube-api-access-c9zjc\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.823127 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-684987dfb8-72476"] Dec 11 08:58:07 crc kubenswrapper[4629]: E1211 08:58:03.823652 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23" containerName="neutron-db-sync" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.823692 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23" containerName="neutron-db-sync" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.823943 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23" containerName="neutron-db-sync" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.824867 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.826516 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.831398 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.831437 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.831923 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-blx9j" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.832141 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.842271 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-684987dfb8-72476"] Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.945357 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-combined-ca-bundle\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.945447 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg4dc\" (UniqueName: \"kubernetes.io/projected/9d2a333e-3183-4709-a7f9-a04d60b90bad-kube-api-access-tg4dc\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.945547 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-public-tls-certs\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.945611 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2a333e-3183-4709-a7f9-a04d60b90bad-logs\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.945683 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-scripts\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.945895 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-config-data\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.945948 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-internal-tls-certs\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.981615 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2thgx" event={"ID":"ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23","Type":"ContainerDied","Data":"684e511da4ac48c50572b92b71b04c9f9cd0d0e7279196909810ddf75cfacab3"} Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.981658 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="684e511da4ac48c50572b92b71b04c9f9cd0d0e7279196909810ddf75cfacab3" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:03.981724 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2thgx" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.047083 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-scripts\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.047190 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-config-data\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.047260 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-internal-tls-certs\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.047290 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-combined-ca-bundle\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.047817 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg4dc\" (UniqueName: \"kubernetes.io/projected/9d2a333e-3183-4709-a7f9-a04d60b90bad-kube-api-access-tg4dc\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.047867 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-public-tls-certs\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.047899 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2a333e-3183-4709-a7f9-a04d60b90bad-logs\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.048276 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2a333e-3183-4709-a7f9-a04d60b90bad-logs\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.050739 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-internal-tls-certs\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.051783 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-scripts\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.052314 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-config-data\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.054336 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-combined-ca-bundle\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.054370 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d2a333e-3183-4709-a7f9-a04d60b90bad-public-tls-certs\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.072303 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg4dc\" (UniqueName: \"kubernetes.io/projected/9d2a333e-3183-4709-a7f9-a04d60b90bad-kube-api-access-tg4dc\") pod \"placement-684987dfb8-72476\" (UID: \"9d2a333e-3183-4709-a7f9-a04d60b90bad\") " pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.144194 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.223606 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp"] Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.223926 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" podUID="297d67cc-431f-45ae-8e51-6df1148414d2" containerName="dnsmasq-dns" containerID="cri-o://f872cc4ce3e0b88b5c46ab7cf729acec7266d8fe8a25f1c94ddfa5d4a60f3b4e" gracePeriod=10 Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.226055 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.273873 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-6bfcl"] Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.279284 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.291056 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-6bfcl"] Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.352440 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.352492 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5mll\" (UniqueName: \"kubernetes.io/projected/a1232ca0-fe10-462c-b663-35eeaa3ded9b-kube-api-access-x5mll\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.352751 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-config\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.352968 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.353003 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.454943 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.455120 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.456111 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.456125 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.456169 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5mll\" (UniqueName: \"kubernetes.io/projected/a1232ca0-fe10-462c-b663-35eeaa3ded9b-kube-api-access-x5mll\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.456311 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-config\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.456429 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.457242 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-config\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.457407 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.476143 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5mll\" (UniqueName: \"kubernetes.io/projected/a1232ca0-fe10-462c-b663-35eeaa3ded9b-kube-api-access-x5mll\") pod \"dnsmasq-dns-5f66db59b9-6bfcl\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.609901 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.728399 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6d446b4c7b-sz524"] Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.729935 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.738014 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.738328 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.741956 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tzz4d" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.744597 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.759079 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6d446b4c7b-sz524"] Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.871338 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-httpd-config\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.871412 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbpfj\" (UniqueName: \"kubernetes.io/projected/91661f71-81be-4498-a6cf-8492a9f194f8-kube-api-access-wbpfj\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.871525 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-combined-ca-bundle\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.871556 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-ovndb-tls-certs\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.871576 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-config\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.973803 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-combined-ca-bundle\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.973907 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-ovndb-tls-certs\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.973936 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-config\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.974050 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-httpd-config\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.974090 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbpfj\" (UniqueName: \"kubernetes.io/projected/91661f71-81be-4498-a6cf-8492a9f194f8-kube-api-access-wbpfj\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.979551 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-httpd-config\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.980010 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-combined-ca-bundle\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.981761 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-config\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:04.983694 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-ovndb-tls-certs\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:05.000958 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbpfj\" (UniqueName: \"kubernetes.io/projected/91661f71-81be-4498-a6cf-8492a9f194f8-kube-api-access-wbpfj\") pod \"neutron-6d446b4c7b-sz524\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:05.070671 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:05.998479 4629 generic.go:334] "Generic (PLEG): container finished" podID="297d67cc-431f-45ae-8e51-6df1148414d2" containerID="f872cc4ce3e0b88b5c46ab7cf729acec7266d8fe8a25f1c94ddfa5d4a60f3b4e" exitCode=0 Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:05.998539 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" event={"ID":"297d67cc-431f-45ae-8e51-6df1148414d2","Type":"ContainerDied","Data":"f872cc4ce3e0b88b5c46ab7cf729acec7266d8fe8a25f1c94ddfa5d4a60f3b4e"} Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.612782 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-59df6557fc-qgbb6"] Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.614945 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.620017 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.621601 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.638141 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59df6557fc-qgbb6"] Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.729291 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-config\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.729351 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-ovndb-tls-certs\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.729405 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-httpd-config\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.729430 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqfzc\" (UniqueName: \"kubernetes.io/projected/1446ae15-0cba-4d28-ae20-0b18ef295ccc-kube-api-access-fqfzc\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.729465 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-combined-ca-bundle\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.729495 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-internal-tls-certs\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.729528 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-public-tls-certs\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.831015 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-httpd-config\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.831081 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqfzc\" (UniqueName: \"kubernetes.io/projected/1446ae15-0cba-4d28-ae20-0b18ef295ccc-kube-api-access-fqfzc\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.831136 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-combined-ca-bundle\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.831191 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-internal-tls-certs\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.831244 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-public-tls-certs\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.831286 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-config\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.831341 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-ovndb-tls-certs\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.839825 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-internal-tls-certs\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.845062 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-ovndb-tls-certs\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.845822 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-combined-ca-bundle\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.846652 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-public-tls-certs\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.846744 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-config\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.859315 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1446ae15-0cba-4d28-ae20-0b18ef295ccc-httpd-config\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.862331 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqfzc\" (UniqueName: \"kubernetes.io/projected/1446ae15-0cba-4d28-ae20-0b18ef295ccc-kube-api-access-fqfzc\") pod \"neutron-59df6557fc-qgbb6\" (UID: \"1446ae15-0cba-4d28-ae20-0b18ef295ccc\") " pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:06.941465 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.609740 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.665521 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-ovsdbserver-sb\") pod \"297d67cc-431f-45ae-8e51-6df1148414d2\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.665647 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-dns-svc\") pod \"297d67cc-431f-45ae-8e51-6df1148414d2\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.665673 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-config\") pod \"297d67cc-431f-45ae-8e51-6df1148414d2\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.665737 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzt9t\" (UniqueName: \"kubernetes.io/projected/297d67cc-431f-45ae-8e51-6df1148414d2-kube-api-access-lzt9t\") pod \"297d67cc-431f-45ae-8e51-6df1148414d2\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.665830 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-ovsdbserver-nb\") pod \"297d67cc-431f-45ae-8e51-6df1148414d2\" (UID: \"297d67cc-431f-45ae-8e51-6df1148414d2\") " Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.681670 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/297d67cc-431f-45ae-8e51-6df1148414d2-kube-api-access-lzt9t" (OuterVolumeSpecName: "kube-api-access-lzt9t") pod "297d67cc-431f-45ae-8e51-6df1148414d2" (UID: "297d67cc-431f-45ae-8e51-6df1148414d2"). InnerVolumeSpecName "kube-api-access-lzt9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.752726 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-config" (OuterVolumeSpecName: "config") pod "297d67cc-431f-45ae-8e51-6df1148414d2" (UID: "297d67cc-431f-45ae-8e51-6df1148414d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.766399 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-684987dfb8-72476"] Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.768608 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.768761 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzt9t\" (UniqueName: \"kubernetes.io/projected/297d67cc-431f-45ae-8e51-6df1148414d2-kube-api-access-lzt9t\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.789468 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "297d67cc-431f-45ae-8e51-6df1148414d2" (UID: "297d67cc-431f-45ae-8e51-6df1148414d2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.816405 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "297d67cc-431f-45ae-8e51-6df1148414d2" (UID: "297d67cc-431f-45ae-8e51-6df1148414d2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.821592 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "297d67cc-431f-45ae-8e51-6df1148414d2" (UID: "297d67cc-431f-45ae-8e51-6df1148414d2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.856124 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5fccc9499c-f4wbp"] Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.870101 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.870147 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.870157 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/297d67cc-431f-45ae-8e51-6df1148414d2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.951321 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-6bfcl"] Dec 11 08:58:07 crc kubenswrapper[4629]: I1211 08:58:07.999148 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6d446b4c7b-sz524"] Dec 11 08:58:08 crc kubenswrapper[4629]: I1211 08:58:08.048021 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-684987dfb8-72476" event={"ID":"9d2a333e-3183-4709-a7f9-a04d60b90bad","Type":"ContainerStarted","Data":"6b5f5917f6e466af30cbdbcc4af897570ea8635cfdea20185a43aef52742c49a"} Dec 11 08:58:08 crc kubenswrapper[4629]: I1211 08:58:08.053251 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d446b4c7b-sz524" event={"ID":"91661f71-81be-4498-a6cf-8492a9f194f8","Type":"ContainerStarted","Data":"faf5cd6cbbd6d8def799c6f610620fcade31c1728bbaa54505b40af203db4dfa"} Dec 11 08:58:08 crc kubenswrapper[4629]: I1211 08:58:08.067237 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" event={"ID":"a1232ca0-fe10-462c-b663-35eeaa3ded9b","Type":"ContainerStarted","Data":"7ed6540ecce1747ce4c1d76c2991eb2b9f11a57ac99ac087396be34cff60a1f4"} Dec 11 08:58:08 crc kubenswrapper[4629]: I1211 08:58:08.086056 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" event={"ID":"297d67cc-431f-45ae-8e51-6df1148414d2","Type":"ContainerDied","Data":"ec64bc9fe42e683608f0d3dcd0de6271ac7addc05defcccc6e245ec725dbadff"} Dec 11 08:58:08 crc kubenswrapper[4629]: I1211 08:58:08.086926 4629 scope.go:117] "RemoveContainer" containerID="f872cc4ce3e0b88b5c46ab7cf729acec7266d8fe8a25f1c94ddfa5d4a60f3b4e" Dec 11 08:58:08 crc kubenswrapper[4629]: I1211 08:58:08.086083 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp" Dec 11 08:58:08 crc kubenswrapper[4629]: I1211 08:58:08.098578 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5fccc9499c-f4wbp" event={"ID":"2998e2d9-9cba-4c82-ac41-9080e89798a3","Type":"ContainerStarted","Data":"ff2b22e8a624a90085f54086c78696c88fc834185238c11ba4df1e4fd3509d55"} Dec 11 08:58:08 crc kubenswrapper[4629]: I1211 08:58:08.200395 4629 scope.go:117] "RemoveContainer" containerID="c095b56340781885392fda07d1d427c63cfe63b8f15e1a258829087e0b6dac89" Dec 11 08:58:08 crc kubenswrapper[4629]: I1211 08:58:08.252522 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp"] Dec 11 08:58:08 crc kubenswrapper[4629]: I1211 08:58:08.261268 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-vmkcp"] Dec 11 08:58:08 crc kubenswrapper[4629]: I1211 08:58:08.848416 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59df6557fc-qgbb6"] Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.109641 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5fccc9499c-f4wbp" event={"ID":"2998e2d9-9cba-4c82-ac41-9080e89798a3","Type":"ContainerStarted","Data":"77c1da10b880fc6f9bc087eb54ead215d882b0c4d42f95e57be211be556985fd"} Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.110688 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.114786 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-684987dfb8-72476" event={"ID":"9d2a333e-3183-4709-a7f9-a04d60b90bad","Type":"ContainerStarted","Data":"1f53cb68cba36c6649c3c7f49b3446e71950d99a03af8e5c7753658c2a84dc72"} Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.117138 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-684987dfb8-72476" event={"ID":"9d2a333e-3183-4709-a7f9-a04d60b90bad","Type":"ContainerStarted","Data":"c4c0a85fdd1f9ff437f666e9fc376cd87249c4cda478761cc79adef00413e8de"} Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.117218 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.117236 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.119714 4629 generic.go:334] "Generic (PLEG): container finished" podID="a1232ca0-fe10-462c-b663-35eeaa3ded9b" containerID="bfbd8f8b272b5665995307cef2eb9e25b10bf401cbf1ec907f334938aed160f6" exitCode=0 Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.119779 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" event={"ID":"a1232ca0-fe10-462c-b663-35eeaa3ded9b","Type":"ContainerDied","Data":"bfbd8f8b272b5665995307cef2eb9e25b10bf401cbf1ec907f334938aed160f6"} Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.127972 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d446b4c7b-sz524" event={"ID":"91661f71-81be-4498-a6cf-8492a9f194f8","Type":"ContainerStarted","Data":"a50bec030cf93971546ff7733d27187f9d11a937132297cb6ebd5d57df5ea551"} Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.128383 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d446b4c7b-sz524" event={"ID":"91661f71-81be-4498-a6cf-8492a9f194f8","Type":"ContainerStarted","Data":"a3856a0078bb38dba5691aa099ffb72005ac104d4237b4bf6bf90d4e2d7626f7"} Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.129142 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.142136 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5fccc9499c-f4wbp" podStartSLOduration=6.142118763 podStartE2EDuration="6.142118763s" podCreationTimestamp="2025-12-11 08:58:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:58:09.129407424 +0000 UTC m=+1277.194823032" watchObservedRunningTime="2025-12-11 08:58:09.142118763 +0000 UTC m=+1277.207534381" Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.142688 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59df6557fc-qgbb6" event={"ID":"1446ae15-0cba-4d28-ae20-0b18ef295ccc","Type":"ContainerStarted","Data":"5163e90db60d51b67093f0abb6a5456fd1378c13c57c278a51a957d276dfb2e5"} Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.218915 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-684987dfb8-72476" podStartSLOduration=6.218890013 podStartE2EDuration="6.218890013s" podCreationTimestamp="2025-12-11 08:58:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:58:09.184146402 +0000 UTC m=+1277.249562020" watchObservedRunningTime="2025-12-11 08:58:09.218890013 +0000 UTC m=+1277.284305631" Dec 11 08:58:09 crc kubenswrapper[4629]: I1211 08:58:09.222408 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6d446b4c7b-sz524" podStartSLOduration=5.222373941 podStartE2EDuration="5.222373941s" podCreationTimestamp="2025-12-11 08:58:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:58:09.210937882 +0000 UTC m=+1277.276353510" watchObservedRunningTime="2025-12-11 08:58:09.222373941 +0000 UTC m=+1277.287789569" Dec 11 08:58:10 crc kubenswrapper[4629]: I1211 08:58:10.172802 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" event={"ID":"a1232ca0-fe10-462c-b663-35eeaa3ded9b","Type":"ContainerStarted","Data":"ed53225b85b8385984870ebc5e5b9d455d5ba75948cc6a1a28815bdf82499838"} Dec 11 08:58:10 crc kubenswrapper[4629]: I1211 08:58:10.173001 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:10 crc kubenswrapper[4629]: I1211 08:58:10.179506 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59df6557fc-qgbb6" event={"ID":"1446ae15-0cba-4d28-ae20-0b18ef295ccc","Type":"ContainerStarted","Data":"d57e18d286973d6aaab839b611ffe78282300844b5b06258eb759b8e48992929"} Dec 11 08:58:10 crc kubenswrapper[4629]: I1211 08:58:10.179550 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59df6557fc-qgbb6" event={"ID":"1446ae15-0cba-4d28-ae20-0b18ef295ccc","Type":"ContainerStarted","Data":"85a392cd219ddce553a0ba5a8667a3cfd9e8093f8458e5be82d0bf406e4cb5f5"} Dec 11 08:58:10 crc kubenswrapper[4629]: I1211 08:58:10.179987 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:10 crc kubenswrapper[4629]: I1211 08:58:10.204334 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" podStartSLOduration=6.204316193 podStartE2EDuration="6.204316193s" podCreationTimestamp="2025-12-11 08:58:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:58:10.200187283 +0000 UTC m=+1278.265602891" watchObservedRunningTime="2025-12-11 08:58:10.204316193 +0000 UTC m=+1278.269731801" Dec 11 08:58:10 crc kubenswrapper[4629]: I1211 08:58:10.214215 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="297d67cc-431f-45ae-8e51-6df1148414d2" path="/var/lib/kubelet/pods/297d67cc-431f-45ae-8e51-6df1148414d2/volumes" Dec 11 08:58:10 crc kubenswrapper[4629]: I1211 08:58:10.229597 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-59df6557fc-qgbb6" podStartSLOduration=4.229574076 podStartE2EDuration="4.229574076s" podCreationTimestamp="2025-12-11 08:58:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:58:10.216508095 +0000 UTC m=+1278.281923713" watchObservedRunningTime="2025-12-11 08:58:10.229574076 +0000 UTC m=+1278.294989684" Dec 11 08:58:14 crc kubenswrapper[4629]: I1211 08:58:14.612155 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:14 crc kubenswrapper[4629]: I1211 08:58:14.679602 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-vdhgz"] Dec 11 08:58:14 crc kubenswrapper[4629]: I1211 08:58:14.679909 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-vdhgz" podUID="6ddc853e-11fc-4d38-b550-8b42d1997c2b" containerName="dnsmasq-dns" containerID="cri-o://7cd6e7708100c30d73d22b66cf3e0b6d2a282c2a01cefc223709bf698b55984c" gracePeriod=10 Dec 11 08:58:15 crc kubenswrapper[4629]: I1211 08:58:15.217980 4629 generic.go:334] "Generic (PLEG): container finished" podID="6ddc853e-11fc-4d38-b550-8b42d1997c2b" containerID="7cd6e7708100c30d73d22b66cf3e0b6d2a282c2a01cefc223709bf698b55984c" exitCode=0 Dec 11 08:58:15 crc kubenswrapper[4629]: I1211 08:58:15.218025 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-vdhgz" event={"ID":"6ddc853e-11fc-4d38-b550-8b42d1997c2b","Type":"ContainerDied","Data":"7cd6e7708100c30d73d22b66cf3e0b6d2a282c2a01cefc223709bf698b55984c"} Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.213793 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.230162 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-vdhgz" event={"ID":"6ddc853e-11fc-4d38-b550-8b42d1997c2b","Type":"ContainerDied","Data":"b1a220e949c42352a487a5a40b2c71c4c720628b8c01114fc8cf91a5e22442ad"} Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.230228 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-vdhgz" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.230236 4629 scope.go:117] "RemoveContainer" containerID="7cd6e7708100c30d73d22b66cf3e0b6d2a282c2a01cefc223709bf698b55984c" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.322242 4629 scope.go:117] "RemoveContainer" containerID="b3c53854322922e062a73b447c888cb0d882bd1f611f87a1ccc01784ba16cca4" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.346987 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-ovsdbserver-nb\") pod \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.347025 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-config\") pod \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.347072 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-dns-svc\") pod \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.347110 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frls5\" (UniqueName: \"kubernetes.io/projected/6ddc853e-11fc-4d38-b550-8b42d1997c2b-kube-api-access-frls5\") pod \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.347147 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-ovsdbserver-sb\") pod \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\" (UID: \"6ddc853e-11fc-4d38-b550-8b42d1997c2b\") " Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.359938 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ddc853e-11fc-4d38-b550-8b42d1997c2b-kube-api-access-frls5" (OuterVolumeSpecName: "kube-api-access-frls5") pod "6ddc853e-11fc-4d38-b550-8b42d1997c2b" (UID: "6ddc853e-11fc-4d38-b550-8b42d1997c2b"). InnerVolumeSpecName "kube-api-access-frls5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.424905 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6ddc853e-11fc-4d38-b550-8b42d1997c2b" (UID: "6ddc853e-11fc-4d38-b550-8b42d1997c2b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.441169 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-config" (OuterVolumeSpecName: "config") pod "6ddc853e-11fc-4d38-b550-8b42d1997c2b" (UID: "6ddc853e-11fc-4d38-b550-8b42d1997c2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.450441 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.450473 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.450486 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frls5\" (UniqueName: \"kubernetes.io/projected/6ddc853e-11fc-4d38-b550-8b42d1997c2b-kube-api-access-frls5\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.451420 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6ddc853e-11fc-4d38-b550-8b42d1997c2b" (UID: "6ddc853e-11fc-4d38-b550-8b42d1997c2b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.461351 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6ddc853e-11fc-4d38-b550-8b42d1997c2b" (UID: "6ddc853e-11fc-4d38-b550-8b42d1997c2b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.551663 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.552006 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ddc853e-11fc-4d38-b550-8b42d1997c2b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:16 crc kubenswrapper[4629]: E1211 08:58:16.604520 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.629418 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-vdhgz"] Dec 11 08:58:16 crc kubenswrapper[4629]: I1211 08:58:16.640435 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-vdhgz"] Dec 11 08:58:17 crc kubenswrapper[4629]: I1211 08:58:17.240339 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a73b4d6-ad71-4ea6-ab69-17d1c581b911","Type":"ContainerStarted","Data":"c5d02d1f287d826d0f0b38903c19ddc5a50e9b6ed886049f20918751c41c5e1e"} Dec 11 08:58:17 crc kubenswrapper[4629]: I1211 08:58:17.240450 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerName="ceilometer-notification-agent" containerID="cri-o://44d6f0deeece43bb1b79e87e7a318cc1e99e9f992cb9d9bf865246d992cc0b60" gracePeriod=30 Dec 11 08:58:17 crc kubenswrapper[4629]: I1211 08:58:17.240494 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerName="sg-core" containerID="cri-o://a03460dd9d0276ab2a93e7fc6d9666994f9e74e8276b474ef8cdc3c6026bfb01" gracePeriod=30 Dec 11 08:58:17 crc kubenswrapper[4629]: I1211 08:58:17.240589 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 08:58:17 crc kubenswrapper[4629]: I1211 08:58:17.240614 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerName="proxy-httpd" containerID="cri-o://c5d02d1f287d826d0f0b38903c19ddc5a50e9b6ed886049f20918751c41c5e1e" gracePeriod=30 Dec 11 08:58:17 crc kubenswrapper[4629]: I1211 08:58:17.243263 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pdwnh" event={"ID":"c198d7a3-15c2-4716-b9a2-4d008076bfdb","Type":"ContainerStarted","Data":"85efd81d390c7a8ea17a19e461389047e42fbec13324724a785a1edcfb045300"} Dec 11 08:58:17 crc kubenswrapper[4629]: I1211 08:58:17.246621 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vr26n" event={"ID":"9c6d6e77-ab1f-494b-9d78-78544b6068c4","Type":"ContainerStarted","Data":"4a234fb7b276826369e032cc2731d56fbbc927b19e0a4e6f85f00953f805f809"} Dec 11 08:58:17 crc kubenswrapper[4629]: I1211 08:58:17.287353 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-vr26n" podStartSLOduration=3.703177352 podStartE2EDuration="51.287334032s" podCreationTimestamp="2025-12-11 08:57:26 +0000 UTC" firstStartedPulling="2025-12-11 08:57:28.740666691 +0000 UTC m=+1236.806082299" lastFinishedPulling="2025-12-11 08:58:16.324823371 +0000 UTC m=+1284.390238979" observedRunningTime="2025-12-11 08:58:17.284427521 +0000 UTC m=+1285.349843139" watchObservedRunningTime="2025-12-11 08:58:17.287334032 +0000 UTC m=+1285.352749640" Dec 11 08:58:17 crc kubenswrapper[4629]: I1211 08:58:17.308013 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-pdwnh" podStartSLOduration=3.654709887 podStartE2EDuration="50.30799168s" podCreationTimestamp="2025-12-11 08:57:27 +0000 UTC" firstStartedPulling="2025-12-11 08:57:29.670907889 +0000 UTC m=+1237.736323497" lastFinishedPulling="2025-12-11 08:58:16.324189692 +0000 UTC m=+1284.389605290" observedRunningTime="2025-12-11 08:58:17.299660659 +0000 UTC m=+1285.365076287" watchObservedRunningTime="2025-12-11 08:58:17.30799168 +0000 UTC m=+1285.373407288" Dec 11 08:58:18 crc kubenswrapper[4629]: I1211 08:58:18.209161 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ddc853e-11fc-4d38-b550-8b42d1997c2b" path="/var/lib/kubelet/pods/6ddc853e-11fc-4d38-b550-8b42d1997c2b/volumes" Dec 11 08:58:18 crc kubenswrapper[4629]: I1211 08:58:18.258387 4629 generic.go:334] "Generic (PLEG): container finished" podID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerID="c5d02d1f287d826d0f0b38903c19ddc5a50e9b6ed886049f20918751c41c5e1e" exitCode=0 Dec 11 08:58:18 crc kubenswrapper[4629]: I1211 08:58:18.258441 4629 generic.go:334] "Generic (PLEG): container finished" podID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerID="a03460dd9d0276ab2a93e7fc6d9666994f9e74e8276b474ef8cdc3c6026bfb01" exitCode=2 Dec 11 08:58:18 crc kubenswrapper[4629]: I1211 08:58:18.258458 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a73b4d6-ad71-4ea6-ab69-17d1c581b911","Type":"ContainerDied","Data":"c5d02d1f287d826d0f0b38903c19ddc5a50e9b6ed886049f20918751c41c5e1e"} Dec 11 08:58:18 crc kubenswrapper[4629]: I1211 08:58:18.258512 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a73b4d6-ad71-4ea6-ab69-17d1c581b911","Type":"ContainerDied","Data":"a03460dd9d0276ab2a93e7fc6d9666994f9e74e8276b474ef8cdc3c6026bfb01"} Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.310326 4629 generic.go:334] "Generic (PLEG): container finished" podID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerID="44d6f0deeece43bb1b79e87e7a318cc1e99e9f992cb9d9bf865246d992cc0b60" exitCode=0 Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.310392 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a73b4d6-ad71-4ea6-ab69-17d1c581b911","Type":"ContainerDied","Data":"44d6f0deeece43bb1b79e87e7a318cc1e99e9f992cb9d9bf865246d992cc0b60"} Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.310954 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a73b4d6-ad71-4ea6-ab69-17d1c581b911","Type":"ContainerDied","Data":"c8440e532dceed0c309959737566b0d8d8dadbd414c6534d0dbfe2a8e7676368"} Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.310967 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8440e532dceed0c309959737566b0d8d8dadbd414c6534d0dbfe2a8e7676368" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.313681 4629 generic.go:334] "Generic (PLEG): container finished" podID="c198d7a3-15c2-4716-b9a2-4d008076bfdb" containerID="85efd81d390c7a8ea17a19e461389047e42fbec13324724a785a1edcfb045300" exitCode=0 Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.313712 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pdwnh" event={"ID":"c198d7a3-15c2-4716-b9a2-4d008076bfdb","Type":"ContainerDied","Data":"85efd81d390c7a8ea17a19e461389047e42fbec13324724a785a1edcfb045300"} Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.322771 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.462032 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-log-httpd\") pod \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.462097 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-config-data\") pod \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.462144 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-run-httpd\") pod \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.462172 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tfbq\" (UniqueName: \"kubernetes.io/projected/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-kube-api-access-7tfbq\") pod \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.462260 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-sg-core-conf-yaml\") pod \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.462326 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-scripts\") pod \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.462355 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-combined-ca-bundle\") pod \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\" (UID: \"1a73b4d6-ad71-4ea6-ab69-17d1c581b911\") " Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.462689 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1a73b4d6-ad71-4ea6-ab69-17d1c581b911" (UID: "1a73b4d6-ad71-4ea6-ab69-17d1c581b911"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.462783 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1a73b4d6-ad71-4ea6-ab69-17d1c581b911" (UID: "1a73b4d6-ad71-4ea6-ab69-17d1c581b911"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.463717 4629 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.463739 4629 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.472083 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-scripts" (OuterVolumeSpecName: "scripts") pod "1a73b4d6-ad71-4ea6-ab69-17d1c581b911" (UID: "1a73b4d6-ad71-4ea6-ab69-17d1c581b911"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.472364 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-kube-api-access-7tfbq" (OuterVolumeSpecName: "kube-api-access-7tfbq") pod "1a73b4d6-ad71-4ea6-ab69-17d1c581b911" (UID: "1a73b4d6-ad71-4ea6-ab69-17d1c581b911"). InnerVolumeSpecName "kube-api-access-7tfbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.490086 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1a73b4d6-ad71-4ea6-ab69-17d1c581b911" (UID: "1a73b4d6-ad71-4ea6-ab69-17d1c581b911"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.511040 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a73b4d6-ad71-4ea6-ab69-17d1c581b911" (UID: "1a73b4d6-ad71-4ea6-ab69-17d1c581b911"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.527782 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-config-data" (OuterVolumeSpecName: "config-data") pod "1a73b4d6-ad71-4ea6-ab69-17d1c581b911" (UID: "1a73b4d6-ad71-4ea6-ab69-17d1c581b911"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.565890 4629 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.566163 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.566174 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.566183 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:22 crc kubenswrapper[4629]: I1211 08:58:22.566192 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tfbq\" (UniqueName: \"kubernetes.io/projected/1a73b4d6-ad71-4ea6-ab69-17d1c581b911-kube-api-access-7tfbq\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.322151 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.383236 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.392218 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.428149 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:58:23 crc kubenswrapper[4629]: E1211 08:58:23.428586 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerName="sg-core" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.428613 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerName="sg-core" Dec 11 08:58:23 crc kubenswrapper[4629]: E1211 08:58:23.428639 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ddc853e-11fc-4d38-b550-8b42d1997c2b" containerName="dnsmasq-dns" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.428648 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ddc853e-11fc-4d38-b550-8b42d1997c2b" containerName="dnsmasq-dns" Dec 11 08:58:23 crc kubenswrapper[4629]: E1211 08:58:23.428662 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerName="ceilometer-notification-agent" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.428669 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerName="ceilometer-notification-agent" Dec 11 08:58:23 crc kubenswrapper[4629]: E1211 08:58:23.428692 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="297d67cc-431f-45ae-8e51-6df1148414d2" containerName="dnsmasq-dns" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.428701 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="297d67cc-431f-45ae-8e51-6df1148414d2" containerName="dnsmasq-dns" Dec 11 08:58:23 crc kubenswrapper[4629]: E1211 08:58:23.428713 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerName="proxy-httpd" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.428720 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerName="proxy-httpd" Dec 11 08:58:23 crc kubenswrapper[4629]: E1211 08:58:23.428736 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ddc853e-11fc-4d38-b550-8b42d1997c2b" containerName="init" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.428743 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ddc853e-11fc-4d38-b550-8b42d1997c2b" containerName="init" Dec 11 08:58:23 crc kubenswrapper[4629]: E1211 08:58:23.428764 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="297d67cc-431f-45ae-8e51-6df1148414d2" containerName="init" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.428775 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="297d67cc-431f-45ae-8e51-6df1148414d2" containerName="init" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.429002 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="297d67cc-431f-45ae-8e51-6df1148414d2" containerName="dnsmasq-dns" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.429022 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ddc853e-11fc-4d38-b550-8b42d1997c2b" containerName="dnsmasq-dns" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.429034 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerName="proxy-httpd" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.429050 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerName="ceilometer-notification-agent" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.429058 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" containerName="sg-core" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.431052 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.436183 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.437816 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.457956 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.585046 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/166a2d96-b521-4037-b8e4-e589b8d24a99-run-httpd\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.585552 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/166a2d96-b521-4037-b8e4-e589b8d24a99-log-httpd\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.585583 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvzwm\" (UniqueName: \"kubernetes.io/projected/166a2d96-b521-4037-b8e4-e589b8d24a99-kube-api-access-xvzwm\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.585622 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-scripts\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.585650 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.585669 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.585725 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-config-data\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.686867 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.686978 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-config-data\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.687068 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/166a2d96-b521-4037-b8e4-e589b8d24a99-run-httpd\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.687114 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/166a2d96-b521-4037-b8e4-e589b8d24a99-log-httpd\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.687133 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvzwm\" (UniqueName: \"kubernetes.io/projected/166a2d96-b521-4037-b8e4-e589b8d24a99-kube-api-access-xvzwm\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.687150 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-scripts\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.687173 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.689396 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/166a2d96-b521-4037-b8e4-e589b8d24a99-run-httpd\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.691665 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/166a2d96-b521-4037-b8e4-e589b8d24a99-log-httpd\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.693603 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-scripts\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.694210 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-config-data\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.695332 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.695340 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.709069 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvzwm\" (UniqueName: \"kubernetes.io/projected/166a2d96-b521-4037-b8e4-e589b8d24a99-kube-api-access-xvzwm\") pod \"ceilometer-0\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.760266 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.761117 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.900910 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c198d7a3-15c2-4716-b9a2-4d008076bfdb-combined-ca-bundle\") pod \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\" (UID: \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\") " Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.900966 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwclc\" (UniqueName: \"kubernetes.io/projected/c198d7a3-15c2-4716-b9a2-4d008076bfdb-kube-api-access-qwclc\") pod \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\" (UID: \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\") " Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.901093 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c198d7a3-15c2-4716-b9a2-4d008076bfdb-db-sync-config-data\") pod \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\" (UID: \"c198d7a3-15c2-4716-b9a2-4d008076bfdb\") " Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.906918 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c198d7a3-15c2-4716-b9a2-4d008076bfdb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c198d7a3-15c2-4716-b9a2-4d008076bfdb" (UID: "c198d7a3-15c2-4716-b9a2-4d008076bfdb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.907816 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c198d7a3-15c2-4716-b9a2-4d008076bfdb-kube-api-access-qwclc" (OuterVolumeSpecName: "kube-api-access-qwclc") pod "c198d7a3-15c2-4716-b9a2-4d008076bfdb" (UID: "c198d7a3-15c2-4716-b9a2-4d008076bfdb"). InnerVolumeSpecName "kube-api-access-qwclc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:58:23 crc kubenswrapper[4629]: I1211 08:58:23.929402 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c198d7a3-15c2-4716-b9a2-4d008076bfdb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c198d7a3-15c2-4716-b9a2-4d008076bfdb" (UID: "c198d7a3-15c2-4716-b9a2-4d008076bfdb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.003815 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c198d7a3-15c2-4716-b9a2-4d008076bfdb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.003898 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwclc\" (UniqueName: \"kubernetes.io/projected/c198d7a3-15c2-4716-b9a2-4d008076bfdb-kube-api-access-qwclc\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.003993 4629 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c198d7a3-15c2-4716-b9a2-4d008076bfdb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:24 crc kubenswrapper[4629]: W1211 08:58:24.200920 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod166a2d96_b521_4037_b8e4_e589b8d24a99.slice/crio-8506948b0318ab9d902bc68203bfcf79b5c55a6505ca82894d32de7d286d541a WatchSource:0}: Error finding container 8506948b0318ab9d902bc68203bfcf79b5c55a6505ca82894d32de7d286d541a: Status 404 returned error can't find the container with id 8506948b0318ab9d902bc68203bfcf79b5c55a6505ca82894d32de7d286d541a Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.215349 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a73b4d6-ad71-4ea6-ab69-17d1c581b911" path="/var/lib/kubelet/pods/1a73b4d6-ad71-4ea6-ab69-17d1c581b911/volumes" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.216176 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.331754 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"166a2d96-b521-4037-b8e4-e589b8d24a99","Type":"ContainerStarted","Data":"8506948b0318ab9d902bc68203bfcf79b5c55a6505ca82894d32de7d286d541a"} Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.333209 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pdwnh" event={"ID":"c198d7a3-15c2-4716-b9a2-4d008076bfdb","Type":"ContainerDied","Data":"763e12b36df5966d1af5a1488deb657a00e7415f864c4da114fcf556763e8b91"} Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.333246 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="763e12b36df5966d1af5a1488deb657a00e7415f864c4da114fcf556763e8b91" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.333252 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pdwnh" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.335151 4629 generic.go:334] "Generic (PLEG): container finished" podID="9c6d6e77-ab1f-494b-9d78-78544b6068c4" containerID="4a234fb7b276826369e032cc2731d56fbbc927b19e0a4e6f85f00953f805f809" exitCode=0 Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.335178 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vr26n" event={"ID":"9c6d6e77-ab1f-494b-9d78-78544b6068c4","Type":"ContainerDied","Data":"4a234fb7b276826369e032cc2731d56fbbc927b19e0a4e6f85f00953f805f809"} Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.652527 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-84ddb5cbd7-jjz7r"] Dec 11 08:58:24 crc kubenswrapper[4629]: E1211 08:58:24.652989 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c198d7a3-15c2-4716-b9a2-4d008076bfdb" containerName="barbican-db-sync" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.653057 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c198d7a3-15c2-4716-b9a2-4d008076bfdb" containerName="barbican-db-sync" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.653320 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="c198d7a3-15c2-4716-b9a2-4d008076bfdb" containerName="barbican-db-sync" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.654355 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.657901 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-k4426" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.658115 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.664049 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.676364 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-695db84c66-bxp2x"] Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.678156 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.682046 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.773674 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-695db84c66-bxp2x"] Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.788056 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-84ddb5cbd7-jjz7r"] Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.816298 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d220d34a-68ce-4217-a15c-4d994b87d988-logs\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.816343 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d220d34a-68ce-4217-a15c-4d994b87d988-config-data\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.816374 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pskhj\" (UniqueName: \"kubernetes.io/projected/3a9fa1b9-acec-4136-9819-d76fa0f88183-kube-api-access-pskhj\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.816396 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d220d34a-68ce-4217-a15c-4d994b87d988-config-data-custom\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.816418 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4ssp\" (UniqueName: \"kubernetes.io/projected/d220d34a-68ce-4217-a15c-4d994b87d988-kube-api-access-d4ssp\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.816486 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a9fa1b9-acec-4136-9819-d76fa0f88183-logs\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.816517 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a9fa1b9-acec-4136-9819-d76fa0f88183-config-data-custom\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.816532 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a9fa1b9-acec-4136-9819-d76fa0f88183-config-data\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.816559 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d220d34a-68ce-4217-a15c-4d994b87d988-combined-ca-bundle\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.816578 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9fa1b9-acec-4136-9819-d76fa0f88183-combined-ca-bundle\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.855810 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-869f779d85-2j94p"] Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.928208 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a9fa1b9-acec-4136-9819-d76fa0f88183-logs\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.944877 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a9fa1b9-acec-4136-9819-d76fa0f88183-config-data-custom\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.945769 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a9fa1b9-acec-4136-9819-d76fa0f88183-config-data\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.946466 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d220d34a-68ce-4217-a15c-4d994b87d988-combined-ca-bundle\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.946537 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9fa1b9-acec-4136-9819-d76fa0f88183-combined-ca-bundle\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.946630 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d220d34a-68ce-4217-a15c-4d994b87d988-logs\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.946667 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d220d34a-68ce-4217-a15c-4d994b87d988-config-data\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.946716 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pskhj\" (UniqueName: \"kubernetes.io/projected/3a9fa1b9-acec-4136-9819-d76fa0f88183-kube-api-access-pskhj\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.946758 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d220d34a-68ce-4217-a15c-4d994b87d988-config-data-custom\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.946796 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4ssp\" (UniqueName: \"kubernetes.io/projected/d220d34a-68ce-4217-a15c-4d994b87d988-kube-api-access-d4ssp\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.929411 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a9fa1b9-acec-4136-9819-d76fa0f88183-logs\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.958572 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-2j94p"] Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.958697 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.960381 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d220d34a-68ce-4217-a15c-4d994b87d988-logs\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.976036 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9fa1b9-acec-4136-9819-d76fa0f88183-combined-ca-bundle\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.977157 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d220d34a-68ce-4217-a15c-4d994b87d988-combined-ca-bundle\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.984596 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a9fa1b9-acec-4136-9819-d76fa0f88183-config-data-custom\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:24 crc kubenswrapper[4629]: I1211 08:58:24.984870 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a9fa1b9-acec-4136-9819-d76fa0f88183-config-data\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.001754 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4ssp\" (UniqueName: \"kubernetes.io/projected/d220d34a-68ce-4217-a15c-4d994b87d988-kube-api-access-d4ssp\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.014663 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d220d34a-68ce-4217-a15c-4d994b87d988-config-data\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.017195 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d220d34a-68ce-4217-a15c-4d994b87d988-config-data-custom\") pod \"barbican-keystone-listener-695db84c66-bxp2x\" (UID: \"d220d34a-68ce-4217-a15c-4d994b87d988\") " pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.017301 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-f69bb6594-gttr7"] Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.020760 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.028555 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pskhj\" (UniqueName: \"kubernetes.io/projected/3a9fa1b9-acec-4136-9819-d76fa0f88183-kube-api-access-pskhj\") pod \"barbican-worker-84ddb5cbd7-jjz7r\" (UID: \"3a9fa1b9-acec-4136-9819-d76fa0f88183\") " pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.028874 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.065155 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.067487 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f69bb6594-gttr7"] Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.082324 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.156932 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz5rj\" (UniqueName: \"kubernetes.io/projected/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-kube-api-access-lz5rj\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.157007 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-config\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.157036 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-combined-ca-bundle\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.157068 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.157095 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-config-data-custom\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.157143 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-dns-svc\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.157205 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-config-data\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.157244 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrl6m\" (UniqueName: \"kubernetes.io/projected/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-kube-api-access-vrl6m\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.157265 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-logs\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.157295 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.259059 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-dns-svc\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.259176 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-config-data\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.259221 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrl6m\" (UniqueName: \"kubernetes.io/projected/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-kube-api-access-vrl6m\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.259236 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-logs\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.259262 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.259376 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz5rj\" (UniqueName: \"kubernetes.io/projected/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-kube-api-access-lz5rj\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.259409 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-config\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.259423 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-combined-ca-bundle\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.259460 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.259477 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-config-data-custom\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.259948 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-dns-svc\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.260505 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.263702 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-logs\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.264496 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-config\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.265087 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.267472 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-combined-ca-bundle\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.270959 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-config-data-custom\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.277978 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-config-data\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.295433 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz5rj\" (UniqueName: \"kubernetes.io/projected/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-kube-api-access-lz5rj\") pod \"barbican-api-f69bb6594-gttr7\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.311640 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrl6m\" (UniqueName: \"kubernetes.io/projected/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-kube-api-access-vrl6m\") pod \"dnsmasq-dns-869f779d85-2j94p\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.390768 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"166a2d96-b521-4037-b8e4-e589b8d24a99","Type":"ContainerStarted","Data":"ba7090b15c4b352ecb368174c39769de7265b8178afb8d4ee1ff40322e920c51"} Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.391254 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.494025 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:25 crc kubenswrapper[4629]: I1211 08:58:25.675885 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-84ddb5cbd7-jjz7r"] Dec 11 08:58:25 crc kubenswrapper[4629]: W1211 08:58:25.717392 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a9fa1b9_acec_4136_9819_d76fa0f88183.slice/crio-c89ecbf224a8cfca810fc3d4fc64bea2c307379c562919e5c49486440fbf7434 WatchSource:0}: Error finding container c89ecbf224a8cfca810fc3d4fc64bea2c307379c562919e5c49486440fbf7434: Status 404 returned error can't find the container with id c89ecbf224a8cfca810fc3d4fc64bea2c307379c562919e5c49486440fbf7434 Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.049957 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-695db84c66-bxp2x"] Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.152456 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vr26n" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.224220 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-config-data\") pod \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.224779 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-combined-ca-bundle\") pod \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.224891 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c6d6e77-ab1f-494b-9d78-78544b6068c4-etc-machine-id\") pod \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.224979 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-db-sync-config-data\") pod \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.225060 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn2jx\" (UniqueName: \"kubernetes.io/projected/9c6d6e77-ab1f-494b-9d78-78544b6068c4-kube-api-access-dn2jx\") pod \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.225201 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-scripts\") pod \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\" (UID: \"9c6d6e77-ab1f-494b-9d78-78544b6068c4\") " Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.227819 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c6d6e77-ab1f-494b-9d78-78544b6068c4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9c6d6e77-ab1f-494b-9d78-78544b6068c4" (UID: "9c6d6e77-ab1f-494b-9d78-78544b6068c4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.236742 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c6d6e77-ab1f-494b-9d78-78544b6068c4-kube-api-access-dn2jx" (OuterVolumeSpecName: "kube-api-access-dn2jx") pod "9c6d6e77-ab1f-494b-9d78-78544b6068c4" (UID: "9c6d6e77-ab1f-494b-9d78-78544b6068c4"). InnerVolumeSpecName "kube-api-access-dn2jx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.255989 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-scripts" (OuterVolumeSpecName: "scripts") pod "9c6d6e77-ab1f-494b-9d78-78544b6068c4" (UID: "9c6d6e77-ab1f-494b-9d78-78544b6068c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.257987 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9c6d6e77-ab1f-494b-9d78-78544b6068c4" (UID: "9c6d6e77-ab1f-494b-9d78-78544b6068c4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.328142 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn2jx\" (UniqueName: \"kubernetes.io/projected/9c6d6e77-ab1f-494b-9d78-78544b6068c4-kube-api-access-dn2jx\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.328456 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.328589 4629 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c6d6e77-ab1f-494b-9d78-78544b6068c4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.328678 4629 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.354009 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c6d6e77-ab1f-494b-9d78-78544b6068c4" (UID: "9c6d6e77-ab1f-494b-9d78-78544b6068c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:26 crc kubenswrapper[4629]: W1211 08:58:26.389958 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod032b7bc2_e4a8_473f_9de9_d222e75d7ce7.slice/crio-aecece0165f4acdc10a5f70b5250f8ecbe9dd1c1922d196863cf48e3a526c3d9 WatchSource:0}: Error finding container aecece0165f4acdc10a5f70b5250f8ecbe9dd1c1922d196863cf48e3a526c3d9: Status 404 returned error can't find the container with id aecece0165f4acdc10a5f70b5250f8ecbe9dd1c1922d196863cf48e3a526c3d9 Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.406920 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-2j94p"] Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.407784 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-2j94p" event={"ID":"032b7bc2-e4a8-473f-9de9-d222e75d7ce7","Type":"ContainerStarted","Data":"aecece0165f4acdc10a5f70b5250f8ecbe9dd1c1922d196863cf48e3a526c3d9"} Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.428480 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f69bb6594-gttr7"] Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.430285 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" event={"ID":"3a9fa1b9-acec-4136-9819-d76fa0f88183","Type":"ContainerStarted","Data":"c89ecbf224a8cfca810fc3d4fc64bea2c307379c562919e5c49486440fbf7434"} Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.434764 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:26 crc kubenswrapper[4629]: W1211 08:58:26.447312 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2dd17bcf_97c4_47f0_86a5_ee28f3a480b6.slice/crio-e845e13b0511fc25d7067f522dddb3f543188107f409fd8c04fc0431e7df2e4b WatchSource:0}: Error finding container e845e13b0511fc25d7067f522dddb3f543188107f409fd8c04fc0431e7df2e4b: Status 404 returned error can't find the container with id e845e13b0511fc25d7067f522dddb3f543188107f409fd8c04fc0431e7df2e4b Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.448229 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vr26n" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.448638 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vr26n" event={"ID":"9c6d6e77-ab1f-494b-9d78-78544b6068c4","Type":"ContainerDied","Data":"123c9f8e939769640846d802f236a8d41baee10da2d99b5eef4f9f3c784ffce3"} Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.448672 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="123c9f8e939769640846d802f236a8d41baee10da2d99b5eef4f9f3c784ffce3" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.450464 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" event={"ID":"d220d34a-68ce-4217-a15c-4d994b87d988","Type":"ContainerStarted","Data":"53e597ac45dad3833464baf76b12f71ae79b2b2b789490686cbc7c98ba276af6"} Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.455541 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-config-data" (OuterVolumeSpecName: "config-data") pod "9c6d6e77-ab1f-494b-9d78-78544b6068c4" (UID: "9c6d6e77-ab1f-494b-9d78-78544b6068c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.536546 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6d6e77-ab1f-494b-9d78-78544b6068c4-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.711881 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:58:26 crc kubenswrapper[4629]: E1211 08:58:26.712697 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c6d6e77-ab1f-494b-9d78-78544b6068c4" containerName="cinder-db-sync" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.712718 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c6d6e77-ab1f-494b-9d78-78544b6068c4" containerName="cinder-db-sync" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.713010 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c6d6e77-ab1f-494b-9d78-78544b6068c4" containerName="cinder-db-sync" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.714228 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.728657 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.748869 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slmf2\" (UniqueName: \"kubernetes.io/projected/015893de-93e3-438c-a52e-35e4fbd64b80-kube-api-access-slmf2\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.748974 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.749009 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/015893de-93e3-438c-a52e-35e4fbd64b80-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.749150 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.749233 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-scripts\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.749356 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-config-data\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.779614 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.852194 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-scripts\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.854114 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-config-data\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.854209 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slmf2\" (UniqueName: \"kubernetes.io/projected/015893de-93e3-438c-a52e-35e4fbd64b80-kube-api-access-slmf2\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.854295 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.854318 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/015893de-93e3-438c-a52e-35e4fbd64b80-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.854491 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.862272 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/015893de-93e3-438c-a52e-35e4fbd64b80-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.865055 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.865517 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-scripts\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.867221 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-config-data\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.873606 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-2j94p"] Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.875677 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.894108 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slmf2\" (UniqueName: \"kubernetes.io/projected/015893de-93e3-438c-a52e-35e4fbd64b80-kube-api-access-slmf2\") pod \"cinder-scheduler-0\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.923924 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.925479 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.928199 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.942154 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-jcpxn"] Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.943754 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.957923 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-config-data\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.957983 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-logs\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.958010 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.958083 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.958103 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-dns-svc\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.958153 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-config\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.958195 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg2xm\" (UniqueName: \"kubernetes.io/projected/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-kube-api-access-jg2xm\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.958223 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-scripts\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.958238 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.958259 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-687zr\" (UniqueName: \"kubernetes.io/projected/a5290303-73ff-47c7-8b4e-6518841d75ec-kube-api-access-687zr\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.958310 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.958353 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-config-data-custom\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.979994 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:58:26 crc kubenswrapper[4629]: I1211 08:58:26.996049 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-jcpxn"] Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.059376 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-logs\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.059725 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.059783 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.059801 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-dns-svc\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.059831 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-config\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.059867 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg2xm\" (UniqueName: \"kubernetes.io/projected/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-kube-api-access-jg2xm\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.059896 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-scripts\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.059911 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.059929 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-687zr\" (UniqueName: \"kubernetes.io/projected/a5290303-73ff-47c7-8b4e-6518841d75ec-kube-api-access-687zr\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.059976 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.060000 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-config-data-custom\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.060036 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-config-data\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.060803 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-config\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.061109 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-logs\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.065588 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.066487 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-dns-svc\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.067307 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.070879 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.071324 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.073987 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-scripts\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.076564 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-config-data-custom\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.077417 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.078773 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-config-data\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.105592 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-687zr\" (UniqueName: \"kubernetes.io/projected/a5290303-73ff-47c7-8b4e-6518841d75ec-kube-api-access-687zr\") pod \"dnsmasq-dns-58db5546cc-jcpxn\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.113516 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg2xm\" (UniqueName: \"kubernetes.io/projected/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-kube-api-access-jg2xm\") pod \"cinder-api-0\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.259828 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.288404 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.478096 4629 generic.go:334] "Generic (PLEG): container finished" podID="032b7bc2-e4a8-473f-9de9-d222e75d7ce7" containerID="10eac373b1d3ca3527972c79a955f707cb99ec6aae12d11597a801587fc1cf84" exitCode=0 Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.478235 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-2j94p" event={"ID":"032b7bc2-e4a8-473f-9de9-d222e75d7ce7","Type":"ContainerDied","Data":"10eac373b1d3ca3527972c79a955f707cb99ec6aae12d11597a801587fc1cf84"} Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.499834 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f69bb6594-gttr7" event={"ID":"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6","Type":"ContainerStarted","Data":"7d8aee31d5cc0d7710661f70f4cd16fd07e6264d02b55c6e9e30518f043c3bac"} Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.499897 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f69bb6594-gttr7" event={"ID":"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6","Type":"ContainerStarted","Data":"e845e13b0511fc25d7067f522dddb3f543188107f409fd8c04fc0431e7df2e4b"} Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.587973 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"166a2d96-b521-4037-b8e4-e589b8d24a99","Type":"ContainerStarted","Data":"5ff949810683ed0ef95a36a8f8ae77c7531378081668c5e067079763eb3bf0d0"} Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.651807 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:58:27 crc kubenswrapper[4629]: I1211 08:58:27.960339 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.029343 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.186635 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-jcpxn"] Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.209283 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrl6m\" (UniqueName: \"kubernetes.io/projected/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-kube-api-access-vrl6m\") pod \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.209339 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-config\") pod \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.209469 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-ovsdbserver-sb\") pod \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.209516 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-ovsdbserver-nb\") pod \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.209603 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-dns-svc\") pod \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\" (UID: \"032b7bc2-e4a8-473f-9de9-d222e75d7ce7\") " Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.242733 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-kube-api-access-vrl6m" (OuterVolumeSpecName: "kube-api-access-vrl6m") pod "032b7bc2-e4a8-473f-9de9-d222e75d7ce7" (UID: "032b7bc2-e4a8-473f-9de9-d222e75d7ce7"). InnerVolumeSpecName "kube-api-access-vrl6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.261359 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "032b7bc2-e4a8-473f-9de9-d222e75d7ce7" (UID: "032b7bc2-e4a8-473f-9de9-d222e75d7ce7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.262132 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-config" (OuterVolumeSpecName: "config") pod "032b7bc2-e4a8-473f-9de9-d222e75d7ce7" (UID: "032b7bc2-e4a8-473f-9de9-d222e75d7ce7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.279857 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "032b7bc2-e4a8-473f-9de9-d222e75d7ce7" (UID: "032b7bc2-e4a8-473f-9de9-d222e75d7ce7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.280338 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "032b7bc2-e4a8-473f-9de9-d222e75d7ce7" (UID: "032b7bc2-e4a8-473f-9de9-d222e75d7ce7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.316933 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrl6m\" (UniqueName: \"kubernetes.io/projected/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-kube-api-access-vrl6m\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.316963 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.316972 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.316981 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.316991 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/032b7bc2-e4a8-473f-9de9-d222e75d7ce7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.601113 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6","Type":"ContainerStarted","Data":"854e683489d5a8877b8b72138903fd0037e313cb84512fe25db1b50f24bc22da"} Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.620461 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-2j94p" event={"ID":"032b7bc2-e4a8-473f-9de9-d222e75d7ce7","Type":"ContainerDied","Data":"aecece0165f4acdc10a5f70b5250f8ecbe9dd1c1922d196863cf48e3a526c3d9"} Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.620541 4629 scope.go:117] "RemoveContainer" containerID="10eac373b1d3ca3527972c79a955f707cb99ec6aae12d11597a801587fc1cf84" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.620709 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-2j94p" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.652188 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f69bb6594-gttr7" event={"ID":"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6","Type":"ContainerStarted","Data":"003b33ac3be999e1aff0c0d9b934346a512a87d1193f1f74b046e59c483c7dda"} Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.652309 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.652391 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.654287 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"015893de-93e3-438c-a52e-35e4fbd64b80","Type":"ContainerStarted","Data":"70753f39bb5fbb8ea65541471d5ec804d2f1088aa0b52c329a6b2dcfc51bdd7a"} Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.737408 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-2j94p"] Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.764595 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-2j94p"] Dec 11 08:58:28 crc kubenswrapper[4629]: I1211 08:58:28.786528 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-f69bb6594-gttr7" podStartSLOduration=4.786500594 podStartE2EDuration="4.786500594s" podCreationTimestamp="2025-12-11 08:58:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:58:28.707265626 +0000 UTC m=+1296.772681244" watchObservedRunningTime="2025-12-11 08:58:28.786500594 +0000 UTC m=+1296.851916202" Dec 11 08:58:29 crc kubenswrapper[4629]: I1211 08:58:29.664198 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" event={"ID":"a5290303-73ff-47c7-8b4e-6518841d75ec","Type":"ContainerStarted","Data":"c868fdbea4ccd1dfe4ed23ddc98bf42e455fa58aadeeb8e42cc56790aa55b7ef"} Dec 11 08:58:29 crc kubenswrapper[4629]: I1211 08:58:29.666770 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6","Type":"ContainerStarted","Data":"cb70e2d6df831b4742034249b5c9ed2e0334f2d83b5a83f79954fac414eef11c"} Dec 11 08:58:30 crc kubenswrapper[4629]: I1211 08:58:30.212453 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="032b7bc2-e4a8-473f-9de9-d222e75d7ce7" path="/var/lib/kubelet/pods/032b7bc2-e4a8-473f-9de9-d222e75d7ce7/volumes" Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.015886 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.702755 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"015893de-93e3-438c-a52e-35e4fbd64b80","Type":"ContainerStarted","Data":"d9806635827801334b3adc8ba82a4e5d638d7180d14e191e460b9a56999e7df5"} Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.713453 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" event={"ID":"d220d34a-68ce-4217-a15c-4d994b87d988","Type":"ContainerStarted","Data":"c03b75f304a1f0cf76d04493eb40abbc079dc5fef5b0c25eb8c5e60346a4d05a"} Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.713720 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" event={"ID":"d220d34a-68ce-4217-a15c-4d994b87d988","Type":"ContainerStarted","Data":"7d6e21ca35a4d23e38f19e80cab3ee5bfdb92a3d81e7b08775609a3dc5b59eb2"} Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.741957 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"166a2d96-b521-4037-b8e4-e589b8d24a99","Type":"ContainerStarted","Data":"3e05cb3826877a5f3a35761571ce305ae016c58a553de9dd3a8624e082bac6db"} Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.744501 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-695db84c66-bxp2x" podStartSLOduration=3.548161555 podStartE2EDuration="7.744483282s" podCreationTimestamp="2025-12-11 08:58:24 +0000 UTC" firstStartedPulling="2025-12-11 08:58:26.074731486 +0000 UTC m=+1294.140147094" lastFinishedPulling="2025-12-11 08:58:30.271053223 +0000 UTC m=+1298.336468821" observedRunningTime="2025-12-11 08:58:31.742299393 +0000 UTC m=+1299.807715001" watchObservedRunningTime="2025-12-11 08:58:31.744483282 +0000 UTC m=+1299.809898890" Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.774943 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6","Type":"ContainerStarted","Data":"71dd9720317ead0c98a604d8a181df5047d2c718d73a9431208fa124c8e6ccbb"} Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.775172 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerName="cinder-api-log" containerID="cri-o://cb70e2d6df831b4742034249b5c9ed2e0334f2d83b5a83f79954fac414eef11c" gracePeriod=30 Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.775439 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.775710 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerName="cinder-api" containerID="cri-o://71dd9720317ead0c98a604d8a181df5047d2c718d73a9431208fa124c8e6ccbb" gracePeriod=30 Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.793327 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" event={"ID":"3a9fa1b9-acec-4136-9819-d76fa0f88183","Type":"ContainerStarted","Data":"d20b9fbe23367ad5dbf00a4fffbf4cb16b7d0375b7f53d40ddbe1d2fba693201"} Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.793379 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" event={"ID":"3a9fa1b9-acec-4136-9819-d76fa0f88183","Type":"ContainerStarted","Data":"ab27eb666a768cc48229ca6d5df5b2903c22ccf835cff11d5f8676635d471d73"} Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.812892 4629 generic.go:334] "Generic (PLEG): container finished" podID="a5290303-73ff-47c7-8b4e-6518841d75ec" containerID="968141be1aa24ffb4d8165374fb571fe86f434479c72211bb48d5b7fcab78fa3" exitCode=0 Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.813005 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" event={"ID":"a5290303-73ff-47c7-8b4e-6518841d75ec","Type":"ContainerDied","Data":"968141be1aa24ffb4d8165374fb571fe86f434479c72211bb48d5b7fcab78fa3"} Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.842537 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.84251966 podStartE2EDuration="5.84251966s" podCreationTimestamp="2025-12-11 08:58:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:58:31.825465305 +0000 UTC m=+1299.890880913" watchObservedRunningTime="2025-12-11 08:58:31.84251966 +0000 UTC m=+1299.907935268" Dec 11 08:58:31 crc kubenswrapper[4629]: I1211 08:58:31.894671 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-84ddb5cbd7-jjz7r" podStartSLOduration=3.394483462 podStartE2EDuration="7.894652037s" podCreationTimestamp="2025-12-11 08:58:24 +0000 UTC" firstStartedPulling="2025-12-11 08:58:25.722208989 +0000 UTC m=+1293.787624597" lastFinishedPulling="2025-12-11 08:58:30.222377564 +0000 UTC m=+1298.287793172" observedRunningTime="2025-12-11 08:58:31.892563492 +0000 UTC m=+1299.957979100" watchObservedRunningTime="2025-12-11 08:58:31.894652037 +0000 UTC m=+1299.960067645" Dec 11 08:58:32 crc kubenswrapper[4629]: I1211 08:58:32.823919 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"015893de-93e3-438c-a52e-35e4fbd64b80","Type":"ContainerStarted","Data":"185de34a4d8206be0d10ab57a08a2789aa2e36b035d5919e8b467721eb186166"} Dec 11 08:58:32 crc kubenswrapper[4629]: I1211 08:58:32.826471 4629 generic.go:334] "Generic (PLEG): container finished" podID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerID="cb70e2d6df831b4742034249b5c9ed2e0334f2d83b5a83f79954fac414eef11c" exitCode=143 Dec 11 08:58:32 crc kubenswrapper[4629]: I1211 08:58:32.826521 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6","Type":"ContainerDied","Data":"cb70e2d6df831b4742034249b5c9ed2e0334f2d83b5a83f79954fac414eef11c"} Dec 11 08:58:32 crc kubenswrapper[4629]: I1211 08:58:32.828897 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" event={"ID":"a5290303-73ff-47c7-8b4e-6518841d75ec","Type":"ContainerStarted","Data":"f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09"} Dec 11 08:58:32 crc kubenswrapper[4629]: I1211 08:58:32.875304 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.222038093 podStartE2EDuration="6.875284645s" podCreationTimestamp="2025-12-11 08:58:26 +0000 UTC" firstStartedPulling="2025-12-11 08:58:27.676911937 +0000 UTC m=+1295.742327555" lastFinishedPulling="2025-12-11 08:58:30.330158499 +0000 UTC m=+1298.395574107" observedRunningTime="2025-12-11 08:58:32.86559028 +0000 UTC m=+1300.931005888" watchObservedRunningTime="2025-12-11 08:58:32.875284645 +0000 UTC m=+1300.940700253" Dec 11 08:58:32 crc kubenswrapper[4629]: I1211 08:58:32.926140 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" podStartSLOduration=6.92611967 podStartE2EDuration="6.92611967s" podCreationTimestamp="2025-12-11 08:58:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:58:32.925400818 +0000 UTC m=+1300.990816446" watchObservedRunningTime="2025-12-11 08:58:32.92611967 +0000 UTC m=+1300.991535288" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.332862 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-56ffd55c5d-fzz6k"] Dec 11 08:58:33 crc kubenswrapper[4629]: E1211 08:58:33.340309 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032b7bc2-e4a8-473f-9de9-d222e75d7ce7" containerName="init" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.340354 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="032b7bc2-e4a8-473f-9de9-d222e75d7ce7" containerName="init" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.340545 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="032b7bc2-e4a8-473f-9de9-d222e75d7ce7" containerName="init" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.341634 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.345831 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.346472 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.348891 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-56ffd55c5d-fzz6k"] Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.517427 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-combined-ca-bundle\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.517483 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-public-tls-certs\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.517512 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-config-data-custom\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.517529 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-internal-tls-certs\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.517594 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-config-data\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.517667 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e730a359-2929-4370-b853-33a09bc7180c-logs\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.517687 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzpxz\" (UniqueName: \"kubernetes.io/projected/e730a359-2929-4370-b853-33a09bc7180c-kube-api-access-bzpxz\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.619497 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e730a359-2929-4370-b853-33a09bc7180c-logs\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.619752 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzpxz\" (UniqueName: \"kubernetes.io/projected/e730a359-2929-4370-b853-33a09bc7180c-kube-api-access-bzpxz\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.619882 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-combined-ca-bundle\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.620502 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-public-tls-certs\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.620636 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-config-data-custom\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.620732 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-internal-tls-certs\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.620918 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-config-data\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.620219 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e730a359-2929-4370-b853-33a09bc7180c-logs\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.629011 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-config-data\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.633676 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-public-tls-certs\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.645593 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-config-data-custom\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.648364 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-combined-ca-bundle\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.658536 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e730a359-2929-4370-b853-33a09bc7180c-internal-tls-certs\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.671877 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzpxz\" (UniqueName: \"kubernetes.io/projected/e730a359-2929-4370-b853-33a09bc7180c-kube-api-access-bzpxz\") pod \"barbican-api-56ffd55c5d-fzz6k\" (UID: \"e730a359-2929-4370-b853-33a09bc7180c\") " pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.672333 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.853064 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"166a2d96-b521-4037-b8e4-e589b8d24a99","Type":"ContainerStarted","Data":"deaea821b52a11c9b38ca06bd27a7931186a8fd0c19fc8961a8754207778765a"} Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.854145 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:33 crc kubenswrapper[4629]: I1211 08:58:33.885178 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.925259435 podStartE2EDuration="10.88514708s" podCreationTimestamp="2025-12-11 08:58:23 +0000 UTC" firstStartedPulling="2025-12-11 08:58:24.204174516 +0000 UTC m=+1292.269590124" lastFinishedPulling="2025-12-11 08:58:33.164062161 +0000 UTC m=+1301.229477769" observedRunningTime="2025-12-11 08:58:33.881376041 +0000 UTC m=+1301.946791649" watchObservedRunningTime="2025-12-11 08:58:33.88514708 +0000 UTC m=+1301.950562708" Dec 11 08:58:34 crc kubenswrapper[4629]: I1211 08:58:34.112479 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-56ffd55c5d-fzz6k"] Dec 11 08:58:34 crc kubenswrapper[4629]: I1211 08:58:34.938874 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56ffd55c5d-fzz6k" event={"ID":"e730a359-2929-4370-b853-33a09bc7180c","Type":"ContainerStarted","Data":"c14a8d1345c9148ee0b9d82d13176b3e923e5f0c1ca364faca6aeff005630f19"} Dec 11 08:58:34 crc kubenswrapper[4629]: I1211 08:58:34.939203 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 08:58:35 crc kubenswrapper[4629]: I1211 08:58:35.165080 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:35 crc kubenswrapper[4629]: I1211 08:58:35.949766 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56ffd55c5d-fzz6k" event={"ID":"e730a359-2929-4370-b853-33a09bc7180c","Type":"ContainerStarted","Data":"c935950351933242901c59c1a4b34babf366e7717c1a4faba7af3b6f796f865f"} Dec 11 08:58:35 crc kubenswrapper[4629]: I1211 08:58:35.950073 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56ffd55c5d-fzz6k" event={"ID":"e730a359-2929-4370-b853-33a09bc7180c","Type":"ContainerStarted","Data":"7cb3739db5d18af8873a7d1c071c0f26f25fc811486361b3a240d613cf5e3ac8"} Dec 11 08:58:35 crc kubenswrapper[4629]: I1211 08:58:35.950101 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:35 crc kubenswrapper[4629]: I1211 08:58:35.950117 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:35 crc kubenswrapper[4629]: I1211 08:58:35.971494 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-56ffd55c5d-fzz6k" podStartSLOduration=2.971461442 podStartE2EDuration="2.971461442s" podCreationTimestamp="2025-12-11 08:58:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:58:35.967530398 +0000 UTC m=+1304.032946006" watchObservedRunningTime="2025-12-11 08:58:35.971461442 +0000 UTC m=+1304.036877050" Dec 11 08:58:36 crc kubenswrapper[4629]: I1211 08:58:36.989512 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-59df6557fc-qgbb6" Dec 11 08:58:37 crc kubenswrapper[4629]: I1211 08:58:37.068770 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6d446b4c7b-sz524"] Dec 11 08:58:37 crc kubenswrapper[4629]: I1211 08:58:37.069028 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6d446b4c7b-sz524" podUID="91661f71-81be-4498-a6cf-8492a9f194f8" containerName="neutron-api" containerID="cri-o://a3856a0078bb38dba5691aa099ffb72005ac104d4237b4bf6bf90d4e2d7626f7" gracePeriod=30 Dec 11 08:58:37 crc kubenswrapper[4629]: I1211 08:58:37.069462 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6d446b4c7b-sz524" podUID="91661f71-81be-4498-a6cf-8492a9f194f8" containerName="neutron-httpd" containerID="cri-o://a50bec030cf93971546ff7733d27187f9d11a937132297cb6ebd5d57df5ea551" gracePeriod=30 Dec 11 08:58:37 crc kubenswrapper[4629]: I1211 08:58:37.074538 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 11 08:58:37 crc kubenswrapper[4629]: I1211 08:58:37.111531 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="015893de-93e3-438c-a52e-35e4fbd64b80" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.147:8080/\": dial tcp 10.217.0.147:8080: connect: connection refused" Dec 11 08:58:37 crc kubenswrapper[4629]: I1211 08:58:37.299891 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 08:58:37 crc kubenswrapper[4629]: I1211 08:58:37.399506 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-6bfcl"] Dec 11 08:58:37 crc kubenswrapper[4629]: I1211 08:58:37.399740 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" podUID="a1232ca0-fe10-462c-b663-35eeaa3ded9b" containerName="dnsmasq-dns" containerID="cri-o://ed53225b85b8385984870ebc5e5b9d455d5ba75948cc6a1a28815bdf82499838" gracePeriod=10 Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.015064 4629 generic.go:334] "Generic (PLEG): container finished" podID="91661f71-81be-4498-a6cf-8492a9f194f8" containerID="a50bec030cf93971546ff7733d27187f9d11a937132297cb6ebd5d57df5ea551" exitCode=0 Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.015477 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d446b4c7b-sz524" event={"ID":"91661f71-81be-4498-a6cf-8492a9f194f8","Type":"ContainerDied","Data":"a50bec030cf93971546ff7733d27187f9d11a937132297cb6ebd5d57df5ea551"} Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.042750 4629 generic.go:334] "Generic (PLEG): container finished" podID="a1232ca0-fe10-462c-b663-35eeaa3ded9b" containerID="ed53225b85b8385984870ebc5e5b9d455d5ba75948cc6a1a28815bdf82499838" exitCode=0 Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.043687 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" event={"ID":"a1232ca0-fe10-462c-b663-35eeaa3ded9b","Type":"ContainerDied","Data":"ed53225b85b8385984870ebc5e5b9d455d5ba75948cc6a1a28815bdf82499838"} Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.183046 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.269367 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-ovsdbserver-nb\") pod \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.269499 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-dns-svc\") pod \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.269594 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-config\") pod \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.269646 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-ovsdbserver-sb\") pod \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.269681 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5mll\" (UniqueName: \"kubernetes.io/projected/a1232ca0-fe10-462c-b663-35eeaa3ded9b-kube-api-access-x5mll\") pod \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\" (UID: \"a1232ca0-fe10-462c-b663-35eeaa3ded9b\") " Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.275229 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1232ca0-fe10-462c-b663-35eeaa3ded9b-kube-api-access-x5mll" (OuterVolumeSpecName: "kube-api-access-x5mll") pod "a1232ca0-fe10-462c-b663-35eeaa3ded9b" (UID: "a1232ca0-fe10-462c-b663-35eeaa3ded9b"). InnerVolumeSpecName "kube-api-access-x5mll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.374121 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5mll\" (UniqueName: \"kubernetes.io/projected/a1232ca0-fe10-462c-b663-35eeaa3ded9b-kube-api-access-x5mll\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.381750 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a1232ca0-fe10-462c-b663-35eeaa3ded9b" (UID: "a1232ca0-fe10-462c-b663-35eeaa3ded9b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.381775 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a1232ca0-fe10-462c-b663-35eeaa3ded9b" (UID: "a1232ca0-fe10-462c-b663-35eeaa3ded9b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.406662 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-config" (OuterVolumeSpecName: "config") pod "a1232ca0-fe10-462c-b663-35eeaa3ded9b" (UID: "a1232ca0-fe10-462c-b663-35eeaa3ded9b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.410671 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a1232ca0-fe10-462c-b663-35eeaa3ded9b" (UID: "a1232ca0-fe10-462c-b663-35eeaa3ded9b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.476084 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.476123 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.476135 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.476143 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1232ca0-fe10-462c-b663-35eeaa3ded9b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:38 crc kubenswrapper[4629]: I1211 08:58:38.949394 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.059864 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" event={"ID":"a1232ca0-fe10-462c-b663-35eeaa3ded9b","Type":"ContainerDied","Data":"7ed6540ecce1747ce4c1d76c2991eb2b9f11a57ac99ac087396be34cff60a1f4"} Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.059948 4629 scope.go:117] "RemoveContainer" containerID="ed53225b85b8385984870ebc5e5b9d455d5ba75948cc6a1a28815bdf82499838" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.060195 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-6bfcl" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.062404 4629 generic.go:334] "Generic (PLEG): container finished" podID="91661f71-81be-4498-a6cf-8492a9f194f8" containerID="a3856a0078bb38dba5691aa099ffb72005ac104d4237b4bf6bf90d4e2d7626f7" exitCode=0 Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.062483 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d446b4c7b-sz524" event={"ID":"91661f71-81be-4498-a6cf-8492a9f194f8","Type":"ContainerDied","Data":"a3856a0078bb38dba5691aa099ffb72005ac104d4237b4bf6bf90d4e2d7626f7"} Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.159476 4629 scope.go:117] "RemoveContainer" containerID="bfbd8f8b272b5665995307cef2eb9e25b10bf401cbf1ec907f334938aed160f6" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.161403 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-684987dfb8-72476" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.309976 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-6bfcl"] Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.328090 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-6bfcl"] Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.454345 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.528626 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-combined-ca-bundle\") pod \"91661f71-81be-4498-a6cf-8492a9f194f8\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.528721 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbpfj\" (UniqueName: \"kubernetes.io/projected/91661f71-81be-4498-a6cf-8492a9f194f8-kube-api-access-wbpfj\") pod \"91661f71-81be-4498-a6cf-8492a9f194f8\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.528798 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-httpd-config\") pod \"91661f71-81be-4498-a6cf-8492a9f194f8\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.528882 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-config\") pod \"91661f71-81be-4498-a6cf-8492a9f194f8\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.528947 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-ovndb-tls-certs\") pod \"91661f71-81be-4498-a6cf-8492a9f194f8\" (UID: \"91661f71-81be-4498-a6cf-8492a9f194f8\") " Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.550286 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "91661f71-81be-4498-a6cf-8492a9f194f8" (UID: "91661f71-81be-4498-a6cf-8492a9f194f8"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.578643 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91661f71-81be-4498-a6cf-8492a9f194f8-kube-api-access-wbpfj" (OuterVolumeSpecName: "kube-api-access-wbpfj") pod "91661f71-81be-4498-a6cf-8492a9f194f8" (UID: "91661f71-81be-4498-a6cf-8492a9f194f8"). InnerVolumeSpecName "kube-api-access-wbpfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.631397 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbpfj\" (UniqueName: \"kubernetes.io/projected/91661f71-81be-4498-a6cf-8492a9f194f8-kube-api-access-wbpfj\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.631428 4629 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.711057 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91661f71-81be-4498-a6cf-8492a9f194f8" (UID: "91661f71-81be-4498-a6cf-8492a9f194f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.716074 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-config" (OuterVolumeSpecName: "config") pod "91661f71-81be-4498-a6cf-8492a9f194f8" (UID: "91661f71-81be-4498-a6cf-8492a9f194f8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.750447 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.750492 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.763022 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "91661f71-81be-4498-a6cf-8492a9f194f8" (UID: "91661f71-81be-4498-a6cf-8492a9f194f8"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.854430 4629 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/91661f71-81be-4498-a6cf-8492a9f194f8-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.971048 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-f69bb6594-gttr7" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:58:39 crc kubenswrapper[4629]: I1211 08:58:39.971071 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-f69bb6594-gttr7" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:58:40 crc kubenswrapper[4629]: I1211 08:58:40.073735 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d446b4c7b-sz524" event={"ID":"91661f71-81be-4498-a6cf-8492a9f194f8","Type":"ContainerDied","Data":"faf5cd6cbbd6d8def799c6f610620fcade31c1728bbaa54505b40af203db4dfa"} Dec 11 08:58:40 crc kubenswrapper[4629]: I1211 08:58:40.074744 4629 scope.go:117] "RemoveContainer" containerID="a50bec030cf93971546ff7733d27187f9d11a937132297cb6ebd5d57df5ea551" Dec 11 08:58:40 crc kubenswrapper[4629]: I1211 08:58:40.074027 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6d446b4c7b-sz524" Dec 11 08:58:40 crc kubenswrapper[4629]: I1211 08:58:40.119955 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6d446b4c7b-sz524"] Dec 11 08:58:40 crc kubenswrapper[4629]: I1211 08:58:40.136569 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6d446b4c7b-sz524"] Dec 11 08:58:40 crc kubenswrapper[4629]: I1211 08:58:40.140603 4629 scope.go:117] "RemoveContainer" containerID="a3856a0078bb38dba5691aa099ffb72005ac104d4237b4bf6bf90d4e2d7626f7" Dec 11 08:58:40 crc kubenswrapper[4629]: I1211 08:58:40.222462 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91661f71-81be-4498-a6cf-8492a9f194f8" path="/var/lib/kubelet/pods/91661f71-81be-4498-a6cf-8492a9f194f8/volumes" Dec 11 08:58:40 crc kubenswrapper[4629]: I1211 08:58:40.223261 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1232ca0-fe10-462c-b663-35eeaa3ded9b" path="/var/lib/kubelet/pods/a1232ca0-fe10-462c-b663-35eeaa3ded9b/volumes" Dec 11 08:58:40 crc kubenswrapper[4629]: I1211 08:58:40.580157 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f69bb6594-gttr7" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:58:40 crc kubenswrapper[4629]: I1211 08:58:40.580322 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f69bb6594-gttr7" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:58:42 crc kubenswrapper[4629]: I1211 08:58:42.305997 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.148:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:58:42 crc kubenswrapper[4629]: I1211 08:58:42.386776 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 11 08:58:42 crc kubenswrapper[4629]: I1211 08:58:42.493932 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:58:42 crc kubenswrapper[4629]: I1211 08:58:42.896961 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5fccc9499c-f4wbp" Dec 11 08:58:43 crc kubenswrapper[4629]: I1211 08:58:43.100327 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="015893de-93e3-438c-a52e-35e4fbd64b80" containerName="cinder-scheduler" containerID="cri-o://d9806635827801334b3adc8ba82a4e5d638d7180d14e191e460b9a56999e7df5" gracePeriod=30 Dec 11 08:58:43 crc kubenswrapper[4629]: I1211 08:58:43.100444 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="015893de-93e3-438c-a52e-35e4fbd64b80" containerName="probe" containerID="cri-o://185de34a4d8206be0d10ab57a08a2789aa2e36b035d5919e8b467721eb186166" gracePeriod=30 Dec 11 08:58:45 crc kubenswrapper[4629]: I1211 08:58:45.119236 4629 generic.go:334] "Generic (PLEG): container finished" podID="015893de-93e3-438c-a52e-35e4fbd64b80" containerID="d9806635827801334b3adc8ba82a4e5d638d7180d14e191e460b9a56999e7df5" exitCode=0 Dec 11 08:58:45 crc kubenswrapper[4629]: I1211 08:58:45.119319 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"015893de-93e3-438c-a52e-35e4fbd64b80","Type":"ContainerDied","Data":"d9806635827801334b3adc8ba82a4e5d638d7180d14e191e460b9a56999e7df5"} Dec 11 08:58:45 crc kubenswrapper[4629]: I1211 08:58:45.500616 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:45 crc kubenswrapper[4629]: I1211 08:58:45.623001 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f69bb6594-gttr7" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:58:45 crc kubenswrapper[4629]: I1211 08:58:45.633256 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.140743 4629 generic.go:334] "Generic (PLEG): container finished" podID="015893de-93e3-438c-a52e-35e4fbd64b80" containerID="185de34a4d8206be0d10ab57a08a2789aa2e36b035d5919e8b467721eb186166" exitCode=0 Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.140807 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"015893de-93e3-438c-a52e-35e4fbd64b80","Type":"ContainerDied","Data":"185de34a4d8206be0d10ab57a08a2789aa2e36b035d5919e8b467721eb186166"} Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.150954 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"015893de-93e3-438c-a52e-35e4fbd64b80","Type":"ContainerDied","Data":"70753f39bb5fbb8ea65541471d5ec804d2f1088aa0b52c329a6b2dcfc51bdd7a"} Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.150972 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70753f39bb5fbb8ea65541471d5ec804d2f1088aa0b52c329a6b2dcfc51bdd7a" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.168148 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.357665 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-scripts\") pod \"015893de-93e3-438c-a52e-35e4fbd64b80\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.357761 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-combined-ca-bundle\") pod \"015893de-93e3-438c-a52e-35e4fbd64b80\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.357786 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/015893de-93e3-438c-a52e-35e4fbd64b80-etc-machine-id\") pod \"015893de-93e3-438c-a52e-35e4fbd64b80\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.358040 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-config-data\") pod \"015893de-93e3-438c-a52e-35e4fbd64b80\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.358064 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-config-data-custom\") pod \"015893de-93e3-438c-a52e-35e4fbd64b80\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.358103 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slmf2\" (UniqueName: \"kubernetes.io/projected/015893de-93e3-438c-a52e-35e4fbd64b80-kube-api-access-slmf2\") pod \"015893de-93e3-438c-a52e-35e4fbd64b80\" (UID: \"015893de-93e3-438c-a52e-35e4fbd64b80\") " Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.362506 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/015893de-93e3-438c-a52e-35e4fbd64b80-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "015893de-93e3-438c-a52e-35e4fbd64b80" (UID: "015893de-93e3-438c-a52e-35e4fbd64b80"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.369279 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-scripts" (OuterVolumeSpecName: "scripts") pod "015893de-93e3-438c-a52e-35e4fbd64b80" (UID: "015893de-93e3-438c-a52e-35e4fbd64b80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.373084 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/015893de-93e3-438c-a52e-35e4fbd64b80-kube-api-access-slmf2" (OuterVolumeSpecName: "kube-api-access-slmf2") pod "015893de-93e3-438c-a52e-35e4fbd64b80" (UID: "015893de-93e3-438c-a52e-35e4fbd64b80"). InnerVolumeSpecName "kube-api-access-slmf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.375986 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "015893de-93e3-438c-a52e-35e4fbd64b80" (UID: "015893de-93e3-438c-a52e-35e4fbd64b80"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.447443 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "015893de-93e3-438c-a52e-35e4fbd64b80" (UID: "015893de-93e3-438c-a52e-35e4fbd64b80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.463683 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slmf2\" (UniqueName: \"kubernetes.io/projected/015893de-93e3-438c-a52e-35e4fbd64b80-kube-api-access-slmf2\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.463724 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.463734 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.463742 4629 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/015893de-93e3-438c-a52e-35e4fbd64b80-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.463751 4629 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.614172 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-config-data" (OuterVolumeSpecName: "config-data") pod "015893de-93e3-438c-a52e-35e4fbd64b80" (UID: "015893de-93e3-438c-a52e-35e4fbd64b80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:58:46 crc kubenswrapper[4629]: I1211 08:58:46.684238 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/015893de-93e3-438c-a52e-35e4fbd64b80-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.038175 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 11 08:58:47 crc kubenswrapper[4629]: E1211 08:58:47.038549 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91661f71-81be-4498-a6cf-8492a9f194f8" containerName="neutron-api" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.038565 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="91661f71-81be-4498-a6cf-8492a9f194f8" containerName="neutron-api" Dec 11 08:58:47 crc kubenswrapper[4629]: E1211 08:58:47.038580 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015893de-93e3-438c-a52e-35e4fbd64b80" containerName="cinder-scheduler" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.038586 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="015893de-93e3-438c-a52e-35e4fbd64b80" containerName="cinder-scheduler" Dec 11 08:58:47 crc kubenswrapper[4629]: E1211 08:58:47.038598 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1232ca0-fe10-462c-b663-35eeaa3ded9b" containerName="init" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.038605 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1232ca0-fe10-462c-b663-35eeaa3ded9b" containerName="init" Dec 11 08:58:47 crc kubenswrapper[4629]: E1211 08:58:47.038620 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91661f71-81be-4498-a6cf-8492a9f194f8" containerName="neutron-httpd" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.038626 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="91661f71-81be-4498-a6cf-8492a9f194f8" containerName="neutron-httpd" Dec 11 08:58:47 crc kubenswrapper[4629]: E1211 08:58:47.038635 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1232ca0-fe10-462c-b663-35eeaa3ded9b" containerName="dnsmasq-dns" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.038641 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1232ca0-fe10-462c-b663-35eeaa3ded9b" containerName="dnsmasq-dns" Dec 11 08:58:47 crc kubenswrapper[4629]: E1211 08:58:47.038659 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015893de-93e3-438c-a52e-35e4fbd64b80" containerName="probe" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.038665 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="015893de-93e3-438c-a52e-35e4fbd64b80" containerName="probe" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.038859 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1232ca0-fe10-462c-b663-35eeaa3ded9b" containerName="dnsmasq-dns" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.038873 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="015893de-93e3-438c-a52e-35e4fbd64b80" containerName="probe" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.038879 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="91661f71-81be-4498-a6cf-8492a9f194f8" containerName="neutron-httpd" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.038897 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="91661f71-81be-4498-a6cf-8492a9f194f8" containerName="neutron-api" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.038921 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="015893de-93e3-438c-a52e-35e4fbd64b80" containerName="cinder-scheduler" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.039474 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.041246 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-f6w2r" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.041540 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.042623 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.053584 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.102740 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9eb2d44d-ec14-4f24-901b-ba24da452c54-openstack-config-secret\") pod \"openstackclient\" (UID: \"9eb2d44d-ec14-4f24-901b-ba24da452c54\") " pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.102812 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb2d44d-ec14-4f24-901b-ba24da452c54-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9eb2d44d-ec14-4f24-901b-ba24da452c54\") " pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.102872 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swdhv\" (UniqueName: \"kubernetes.io/projected/9eb2d44d-ec14-4f24-901b-ba24da452c54-kube-api-access-swdhv\") pod \"openstackclient\" (UID: \"9eb2d44d-ec14-4f24-901b-ba24da452c54\") " pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.102930 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9eb2d44d-ec14-4f24-901b-ba24da452c54-openstack-config\") pod \"openstackclient\" (UID: \"9eb2d44d-ec14-4f24-901b-ba24da452c54\") " pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.157441 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.193930 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.204673 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb2d44d-ec14-4f24-901b-ba24da452c54-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9eb2d44d-ec14-4f24-901b-ba24da452c54\") " pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.204792 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swdhv\" (UniqueName: \"kubernetes.io/projected/9eb2d44d-ec14-4f24-901b-ba24da452c54-kube-api-access-swdhv\") pod \"openstackclient\" (UID: \"9eb2d44d-ec14-4f24-901b-ba24da452c54\") " pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.205024 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9eb2d44d-ec14-4f24-901b-ba24da452c54-openstack-config\") pod \"openstackclient\" (UID: \"9eb2d44d-ec14-4f24-901b-ba24da452c54\") " pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.205151 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9eb2d44d-ec14-4f24-901b-ba24da452c54-openstack-config-secret\") pod \"openstackclient\" (UID: \"9eb2d44d-ec14-4f24-901b-ba24da452c54\") " pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.206563 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9eb2d44d-ec14-4f24-901b-ba24da452c54-openstack-config\") pod \"openstackclient\" (UID: \"9eb2d44d-ec14-4f24-901b-ba24da452c54\") " pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.207202 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.209401 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb2d44d-ec14-4f24-901b-ba24da452c54-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9eb2d44d-ec14-4f24-901b-ba24da452c54\") " pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.225981 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9eb2d44d-ec14-4f24-901b-ba24da452c54-openstack-config-secret\") pod \"openstackclient\" (UID: \"9eb2d44d-ec14-4f24-901b-ba24da452c54\") " pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.233429 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.239005 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.239730 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swdhv\" (UniqueName: \"kubernetes.io/projected/9eb2d44d-ec14-4f24-901b-ba24da452c54-kube-api-access-swdhv\") pod \"openstackclient\" (UID: \"9eb2d44d-ec14-4f24-901b-ba24da452c54\") " pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.242591 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.257437 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.318477 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/648e881c-c342-458f-8acd-71f7e1bff9a0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.318566 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/648e881c-c342-458f-8acd-71f7e1bff9a0-config-data\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.318584 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/648e881c-c342-458f-8acd-71f7e1bff9a0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.318620 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/648e881c-c342-458f-8acd-71f7e1bff9a0-scripts\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.318669 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/648e881c-c342-458f-8acd-71f7e1bff9a0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.318689 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2dtf\" (UniqueName: \"kubernetes.io/projected/648e881c-c342-458f-8acd-71f7e1bff9a0-kube-api-access-l2dtf\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.356146 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.148:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.410335 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.419711 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/648e881c-c342-458f-8acd-71f7e1bff9a0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.419809 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/648e881c-c342-458f-8acd-71f7e1bff9a0-config-data\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.419840 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/648e881c-c342-458f-8acd-71f7e1bff9a0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.419912 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/648e881c-c342-458f-8acd-71f7e1bff9a0-scripts\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.419945 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/648e881c-c342-458f-8acd-71f7e1bff9a0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.419974 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2dtf\" (UniqueName: \"kubernetes.io/projected/648e881c-c342-458f-8acd-71f7e1bff9a0-kube-api-access-l2dtf\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.420498 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/648e881c-c342-458f-8acd-71f7e1bff9a0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.453451 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/648e881c-c342-458f-8acd-71f7e1bff9a0-scripts\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.456507 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/648e881c-c342-458f-8acd-71f7e1bff9a0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.458458 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/648e881c-c342-458f-8acd-71f7e1bff9a0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.469270 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/648e881c-c342-458f-8acd-71f7e1bff9a0-config-data\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.472366 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2dtf\" (UniqueName: \"kubernetes.io/projected/648e881c-c342-458f-8acd-71f7e1bff9a0-kube-api-access-l2dtf\") pod \"cinder-scheduler-0\" (UID: \"648e881c-c342-458f-8acd-71f7e1bff9a0\") " pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.614302 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.687072 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-56ffd55c5d-fzz6k" podUID="e730a359-2929-4370-b853-33a09bc7180c" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.150:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 08:58:47 crc kubenswrapper[4629]: I1211 08:58:47.687915 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-56ffd55c5d-fzz6k" podUID="e730a359-2929-4370-b853-33a09bc7180c" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.150:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 08:58:48 crc kubenswrapper[4629]: I1211 08:58:48.145924 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 11 08:58:48 crc kubenswrapper[4629]: I1211 08:58:48.184591 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9eb2d44d-ec14-4f24-901b-ba24da452c54","Type":"ContainerStarted","Data":"2abcde9554cda406558b4d6baf8d7bc726912f2d6c42aa62dd2ba98bf085b02e"} Dec 11 08:58:48 crc kubenswrapper[4629]: I1211 08:58:48.218354 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="015893de-93e3-438c-a52e-35e4fbd64b80" path="/var/lib/kubelet/pods/015893de-93e3-438c-a52e-35e4fbd64b80/volumes" Dec 11 08:58:48 crc kubenswrapper[4629]: I1211 08:58:48.352903 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:58:48 crc kubenswrapper[4629]: W1211 08:58:48.369140 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod648e881c_c342_458f_8acd_71f7e1bff9a0.slice/crio-c31ba5a299f7858dae6786cdbe0d085924f78d6d2475423b8f9a37ca7947c734 WatchSource:0}: Error finding container c31ba5a299f7858dae6786cdbe0d085924f78d6d2475423b8f9a37ca7947c734: Status 404 returned error can't find the container with id c31ba5a299f7858dae6786cdbe0d085924f78d6d2475423b8f9a37ca7947c734 Dec 11 08:58:48 crc kubenswrapper[4629]: I1211 08:58:48.684063 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-56ffd55c5d-fzz6k" podUID="e730a359-2929-4370-b853-33a09bc7180c" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.150:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 08:58:48 crc kubenswrapper[4629]: I1211 08:58:48.684084 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-56ffd55c5d-fzz6k" podUID="e730a359-2929-4370-b853-33a09bc7180c" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.150:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 08:58:49 crc kubenswrapper[4629]: I1211 08:58:49.204266 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"648e881c-c342-458f-8acd-71f7e1bff9a0","Type":"ContainerStarted","Data":"c31ba5a299f7858dae6786cdbe0d085924f78d6d2475423b8f9a37ca7947c734"} Dec 11 08:58:50 crc kubenswrapper[4629]: I1211 08:58:50.237308 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"648e881c-c342-458f-8acd-71f7e1bff9a0","Type":"ContainerStarted","Data":"23eaee18d3af2f8a164ecc6bd86cd42d0231e294fef182af71473184fd386016"} Dec 11 08:58:50 crc kubenswrapper[4629]: I1211 08:58:50.237673 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"648e881c-c342-458f-8acd-71f7e1bff9a0","Type":"ContainerStarted","Data":"abda961d9dacf011fbc512d4733310db204cffc6873d4f19e7ecc16a22357975"} Dec 11 08:58:50 crc kubenswrapper[4629]: I1211 08:58:50.267777 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.267756995 podStartE2EDuration="3.267756995s" podCreationTimestamp="2025-12-11 08:58:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:58:50.265081211 +0000 UTC m=+1318.330496819" watchObservedRunningTime="2025-12-11 08:58:50.267756995 +0000 UTC m=+1318.333172603" Dec 11 08:58:51 crc kubenswrapper[4629]: I1211 08:58:51.733964 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:52 crc kubenswrapper[4629]: I1211 08:58:52.004300 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 11 08:58:52 crc kubenswrapper[4629]: I1211 08:58:52.340055 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-56ffd55c5d-fzz6k" Dec 11 08:58:52 crc kubenswrapper[4629]: I1211 08:58:52.410471 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-f69bb6594-gttr7"] Dec 11 08:58:52 crc kubenswrapper[4629]: I1211 08:58:52.410859 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-f69bb6594-gttr7" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api-log" containerID="cri-o://7d8aee31d5cc0d7710661f70f4cd16fd07e6264d02b55c6e9e30518f043c3bac" gracePeriod=30 Dec 11 08:58:52 crc kubenswrapper[4629]: I1211 08:58:52.410982 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-f69bb6594-gttr7" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api" containerID="cri-o://003b33ac3be999e1aff0c0d9b934346a512a87d1193f1f74b046e59c483c7dda" gracePeriod=30 Dec 11 08:58:52 crc kubenswrapper[4629]: I1211 08:58:52.615311 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 11 08:58:53 crc kubenswrapper[4629]: I1211 08:58:53.278326 4629 generic.go:334] "Generic (PLEG): container finished" podID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerID="7d8aee31d5cc0d7710661f70f4cd16fd07e6264d02b55c6e9e30518f043c3bac" exitCode=143 Dec 11 08:58:53 crc kubenswrapper[4629]: I1211 08:58:53.278402 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f69bb6594-gttr7" event={"ID":"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6","Type":"ContainerDied","Data":"7d8aee31d5cc0d7710661f70f4cd16fd07e6264d02b55c6e9e30518f043c3bac"} Dec 11 08:58:53 crc kubenswrapper[4629]: I1211 08:58:53.770459 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 11 08:58:55 crc kubenswrapper[4629]: I1211 08:58:55.733607 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f69bb6594-gttr7" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": read tcp 10.217.0.2:50936->10.217.0.146:9311: read: connection reset by peer" Dec 11 08:58:55 crc kubenswrapper[4629]: I1211 08:58:55.733649 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f69bb6594-gttr7" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": read tcp 10.217.0.2:50944->10.217.0.146:9311: read: connection reset by peer" Dec 11 08:58:56 crc kubenswrapper[4629]: I1211 08:58:56.318868 4629 generic.go:334] "Generic (PLEG): container finished" podID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerID="003b33ac3be999e1aff0c0d9b934346a512a87d1193f1f74b046e59c483c7dda" exitCode=0 Dec 11 08:58:56 crc kubenswrapper[4629]: I1211 08:58:56.318902 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f69bb6594-gttr7" event={"ID":"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6","Type":"ContainerDied","Data":"003b33ac3be999e1aff0c0d9b934346a512a87d1193f1f74b046e59c483c7dda"} Dec 11 08:58:57 crc kubenswrapper[4629]: I1211 08:58:57.920673 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 11 08:58:58 crc kubenswrapper[4629]: I1211 08:58:58.060552 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:58:58 crc kubenswrapper[4629]: I1211 08:58:58.061398 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="ceilometer-central-agent" containerID="cri-o://ba7090b15c4b352ecb368174c39769de7265b8178afb8d4ee1ff40322e920c51" gracePeriod=30 Dec 11 08:58:58 crc kubenswrapper[4629]: I1211 08:58:58.061425 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="proxy-httpd" containerID="cri-o://deaea821b52a11c9b38ca06bd27a7931186a8fd0c19fc8961a8754207778765a" gracePeriod=30 Dec 11 08:58:58 crc kubenswrapper[4629]: I1211 08:58:58.061460 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="sg-core" containerID="cri-o://3e05cb3826877a5f3a35761571ce305ae016c58a553de9dd3a8624e082bac6db" gracePeriod=30 Dec 11 08:58:58 crc kubenswrapper[4629]: I1211 08:58:58.061544 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="ceilometer-notification-agent" containerID="cri-o://5ff949810683ed0ef95a36a8f8ae77c7531378081668c5e067079763eb3bf0d0" gracePeriod=30 Dec 11 08:58:58 crc kubenswrapper[4629]: I1211 08:58:58.344879 4629 generic.go:334] "Generic (PLEG): container finished" podID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerID="deaea821b52a11c9b38ca06bd27a7931186a8fd0c19fc8961a8754207778765a" exitCode=0 Dec 11 08:58:58 crc kubenswrapper[4629]: I1211 08:58:58.345251 4629 generic.go:334] "Generic (PLEG): container finished" podID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerID="3e05cb3826877a5f3a35761571ce305ae016c58a553de9dd3a8624e082bac6db" exitCode=2 Dec 11 08:58:58 crc kubenswrapper[4629]: I1211 08:58:58.345032 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"166a2d96-b521-4037-b8e4-e589b8d24a99","Type":"ContainerDied","Data":"deaea821b52a11c9b38ca06bd27a7931186a8fd0c19fc8961a8754207778765a"} Dec 11 08:58:58 crc kubenswrapper[4629]: I1211 08:58:58.345288 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"166a2d96-b521-4037-b8e4-e589b8d24a99","Type":"ContainerDied","Data":"3e05cb3826877a5f3a35761571ce305ae016c58a553de9dd3a8624e082bac6db"} Dec 11 08:58:59 crc kubenswrapper[4629]: I1211 08:58:59.371952 4629 generic.go:334] "Generic (PLEG): container finished" podID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerID="ba7090b15c4b352ecb368174c39769de7265b8178afb8d4ee1ff40322e920c51" exitCode=0 Dec 11 08:58:59 crc kubenswrapper[4629]: I1211 08:58:59.372007 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"166a2d96-b521-4037-b8e4-e589b8d24a99","Type":"ContainerDied","Data":"ba7090b15c4b352ecb368174c39769de7265b8178afb8d4ee1ff40322e920c51"} Dec 11 08:59:00 crc kubenswrapper[4629]: I1211 08:59:00.385087 4629 generic.go:334] "Generic (PLEG): container finished" podID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerID="5ff949810683ed0ef95a36a8f8ae77c7531378081668c5e067079763eb3bf0d0" exitCode=0 Dec 11 08:59:00 crc kubenswrapper[4629]: I1211 08:59:00.385154 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"166a2d96-b521-4037-b8e4-e589b8d24a99","Type":"ContainerDied","Data":"5ff949810683ed0ef95a36a8f8ae77c7531378081668c5e067079763eb3bf0d0"} Dec 11 08:59:00 crc kubenswrapper[4629]: I1211 08:59:00.496153 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f69bb6594-gttr7" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": dial tcp 10.217.0.146:9311: connect: connection refused" Dec 11 08:59:00 crc kubenswrapper[4629]: I1211 08:59:00.496279 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f69bb6594-gttr7" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": dial tcp 10.217.0.146:9311: connect: connection refused" Dec 11 08:59:02 crc kubenswrapper[4629]: I1211 08:59:02.262584 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.148:8776/healthcheck\": dial tcp 10.217.0.148:8776: connect: connection refused" Dec 11 08:59:02 crc kubenswrapper[4629]: I1211 08:59:02.411727 4629 generic.go:334] "Generic (PLEG): container finished" podID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerID="71dd9720317ead0c98a604d8a181df5047d2c718d73a9431208fa124c8e6ccbb" exitCode=137 Dec 11 08:59:02 crc kubenswrapper[4629]: I1211 08:59:02.411776 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6","Type":"ContainerDied","Data":"71dd9720317ead0c98a604d8a181df5047d2c718d73a9431208fa124c8e6ccbb"} Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.379817 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.461590 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"166a2d96-b521-4037-b8e4-e589b8d24a99","Type":"ContainerDied","Data":"8506948b0318ab9d902bc68203bfcf79b5c55a6505ca82894d32de7d286d541a"} Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.462000 4629 scope.go:117] "RemoveContainer" containerID="deaea821b52a11c9b38ca06bd27a7931186a8fd0c19fc8961a8754207778765a" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.461839 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.467511 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9eb2d44d-ec14-4f24-901b-ba24da452c54","Type":"ContainerStarted","Data":"3d91b1c1dc425f6f697e4805f66fa94e441e6296980370d5544d800d2358bbbd"} Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.476204 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-config-data\") pod \"166a2d96-b521-4037-b8e4-e589b8d24a99\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.476286 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-scripts\") pod \"166a2d96-b521-4037-b8e4-e589b8d24a99\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.479557 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/166a2d96-b521-4037-b8e4-e589b8d24a99-log-httpd\") pod \"166a2d96-b521-4037-b8e4-e589b8d24a99\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.479676 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/166a2d96-b521-4037-b8e4-e589b8d24a99-run-httpd\") pod \"166a2d96-b521-4037-b8e4-e589b8d24a99\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.480063 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/166a2d96-b521-4037-b8e4-e589b8d24a99-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "166a2d96-b521-4037-b8e4-e589b8d24a99" (UID: "166a2d96-b521-4037-b8e4-e589b8d24a99"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.480676 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/166a2d96-b521-4037-b8e4-e589b8d24a99-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "166a2d96-b521-4037-b8e4-e589b8d24a99" (UID: "166a2d96-b521-4037-b8e4-e589b8d24a99"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.481739 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-sg-core-conf-yaml\") pod \"166a2d96-b521-4037-b8e4-e589b8d24a99\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.482202 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvzwm\" (UniqueName: \"kubernetes.io/projected/166a2d96-b521-4037-b8e4-e589b8d24a99-kube-api-access-xvzwm\") pod \"166a2d96-b521-4037-b8e4-e589b8d24a99\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.482271 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-combined-ca-bundle\") pod \"166a2d96-b521-4037-b8e4-e589b8d24a99\" (UID: \"166a2d96-b521-4037-b8e4-e589b8d24a99\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.483285 4629 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/166a2d96-b521-4037-b8e4-e589b8d24a99-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.483314 4629 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/166a2d96-b521-4037-b8e4-e589b8d24a99-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.508341 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-scripts" (OuterVolumeSpecName: "scripts") pod "166a2d96-b521-4037-b8e4-e589b8d24a99" (UID: "166a2d96-b521-4037-b8e4-e589b8d24a99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.510579 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.614727186 podStartE2EDuration="16.510554583s" podCreationTimestamp="2025-12-11 08:58:47 +0000 UTC" firstStartedPulling="2025-12-11 08:58:48.15841604 +0000 UTC m=+1316.223831648" lastFinishedPulling="2025-12-11 08:59:03.054243437 +0000 UTC m=+1331.119659045" observedRunningTime="2025-12-11 08:59:03.498710921 +0000 UTC m=+1331.564126529" watchObservedRunningTime="2025-12-11 08:59:03.510554583 +0000 UTC m=+1331.575970191" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.512502 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/166a2d96-b521-4037-b8e4-e589b8d24a99-kube-api-access-xvzwm" (OuterVolumeSpecName: "kube-api-access-xvzwm") pod "166a2d96-b521-4037-b8e4-e589b8d24a99" (UID: "166a2d96-b521-4037-b8e4-e589b8d24a99"). InnerVolumeSpecName "kube-api-access-xvzwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.518940 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.536070 4629 scope.go:117] "RemoveContainer" containerID="3e05cb3826877a5f3a35761571ce305ae016c58a553de9dd3a8624e082bac6db" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.545954 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.584098 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-config-data\") pod \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.584152 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-combined-ca-bundle\") pod \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.584179 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-config-data-custom\") pod \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.584227 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz5rj\" (UniqueName: \"kubernetes.io/projected/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-kube-api-access-lz5rj\") pod \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.584248 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-logs\") pod \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\" (UID: \"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.585111 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvzwm\" (UniqueName: \"kubernetes.io/projected/166a2d96-b521-4037-b8e4-e589b8d24a99-kube-api-access-xvzwm\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.585135 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.585199 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-logs" (OuterVolumeSpecName: "logs") pod "2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" (UID: "2dd17bcf-97c4-47f0-86a5-ee28f3a480b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.595195 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" (UID: "2dd17bcf-97c4-47f0-86a5-ee28f3a480b6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.598106 4629 scope.go:117] "RemoveContainer" containerID="5ff949810683ed0ef95a36a8f8ae77c7531378081668c5e067079763eb3bf0d0" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.609752 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-kube-api-access-lz5rj" (OuterVolumeSpecName: "kube-api-access-lz5rj") pod "2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" (UID: "2dd17bcf-97c4-47f0-86a5-ee28f3a480b6"). InnerVolumeSpecName "kube-api-access-lz5rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.625023 4629 scope.go:117] "RemoveContainer" containerID="ba7090b15c4b352ecb368174c39769de7265b8178afb8d4ee1ff40322e920c51" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.644096 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "166a2d96-b521-4037-b8e4-e589b8d24a99" (UID: "166a2d96-b521-4037-b8e4-e589b8d24a99"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.690216 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-logs\") pod \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.690300 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg2xm\" (UniqueName: \"kubernetes.io/projected/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-kube-api-access-jg2xm\") pod \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.690371 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-etc-machine-id\") pod \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.690427 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-combined-ca-bundle\") pod \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.690444 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-config-data-custom\") pod \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.690462 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-config-data\") pod \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.690528 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-scripts\") pod \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\" (UID: \"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6\") " Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.690906 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz5rj\" (UniqueName: \"kubernetes.io/projected/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-kube-api-access-lz5rj\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.690923 4629 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.690922 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" (UID: "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.690933 4629 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.690962 4629 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.691525 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-logs" (OuterVolumeSpecName: "logs") pod "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" (UID: "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.694496 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" (UID: "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.695697 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-scripts" (OuterVolumeSpecName: "scripts") pod "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" (UID: "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.714242 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-kube-api-access-jg2xm" (OuterVolumeSpecName: "kube-api-access-jg2xm") pod "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" (UID: "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6"). InnerVolumeSpecName "kube-api-access-jg2xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.742865 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" (UID: "2dd17bcf-97c4-47f0-86a5-ee28f3a480b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.748929 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-config-data" (OuterVolumeSpecName: "config-data") pod "2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" (UID: "2dd17bcf-97c4-47f0-86a5-ee28f3a480b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.767238 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "166a2d96-b521-4037-b8e4-e589b8d24a99" (UID: "166a2d96-b521-4037-b8e4-e589b8d24a99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.776099 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" (UID: "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.803495 4629 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.803529 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.803539 4629 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.803548 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.803557 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.803568 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.803579 4629 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.803588 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.803598 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg2xm\" (UniqueName: \"kubernetes.io/projected/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-kube-api-access-jg2xm\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.814813 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-config-data" (OuterVolumeSpecName: "config-data") pod "166a2d96-b521-4037-b8e4-e589b8d24a99" (UID: "166a2d96-b521-4037-b8e4-e589b8d24a99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.825467 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-config-data" (OuterVolumeSpecName: "config-data") pod "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" (UID: "ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.904992 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166a2d96-b521-4037-b8e4-e589b8d24a99-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:03 crc kubenswrapper[4629]: I1211 08:59:03.905028 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.095220 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.102336 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.137937 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:59:04 crc kubenswrapper[4629]: E1211 08:59:04.138306 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerName="cinder-api-log" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138334 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerName="cinder-api-log" Dec 11 08:59:04 crc kubenswrapper[4629]: E1211 08:59:04.138349 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="sg-core" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138356 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="sg-core" Dec 11 08:59:04 crc kubenswrapper[4629]: E1211 08:59:04.138363 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="ceilometer-notification-agent" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138369 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="ceilometer-notification-agent" Dec 11 08:59:04 crc kubenswrapper[4629]: E1211 08:59:04.138375 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerName="cinder-api" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138381 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerName="cinder-api" Dec 11 08:59:04 crc kubenswrapper[4629]: E1211 08:59:04.138391 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api-log" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138398 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api-log" Dec 11 08:59:04 crc kubenswrapper[4629]: E1211 08:59:04.138430 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="proxy-httpd" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138440 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="proxy-httpd" Dec 11 08:59:04 crc kubenswrapper[4629]: E1211 08:59:04.138455 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138464 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api" Dec 11 08:59:04 crc kubenswrapper[4629]: E1211 08:59:04.138481 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="ceilometer-central-agent" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138490 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="ceilometer-central-agent" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138663 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="sg-core" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138677 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138689 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="ceilometer-notification-agent" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138699 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerName="cinder-api" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138713 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" containerName="barbican-api-log" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138727 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" containerName="cinder-api-log" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138738 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="proxy-httpd" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.138749 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" containerName="ceilometer-central-agent" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.144232 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.147361 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.147590 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.164286 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.213092 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="166a2d96-b521-4037-b8e4-e589b8d24a99" path="/var/lib/kubelet/pods/166a2d96-b521-4037-b8e4-e589b8d24a99/volumes" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.313375 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-config-data\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.313521 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a515b487-9238-4378-82b6-f190dfc360f7-run-httpd\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.313550 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.313569 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.314020 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-scripts\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.314139 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9khrn\" (UniqueName: \"kubernetes.io/projected/a515b487-9238-4378-82b6-f190dfc360f7-kube-api-access-9khrn\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.314246 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a515b487-9238-4378-82b6-f190dfc360f7-log-httpd\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.416187 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-scripts\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.416255 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9khrn\" (UniqueName: \"kubernetes.io/projected/a515b487-9238-4378-82b6-f190dfc360f7-kube-api-access-9khrn\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.416295 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a515b487-9238-4378-82b6-f190dfc360f7-log-httpd\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.416332 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-config-data\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.416374 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a515b487-9238-4378-82b6-f190dfc360f7-run-httpd\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.416391 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.416406 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.416929 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a515b487-9238-4378-82b6-f190dfc360f7-log-httpd\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.417222 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a515b487-9238-4378-82b6-f190dfc360f7-run-httpd\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.422701 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-scripts\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.424539 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-config-data\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.429820 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.433042 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.443584 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9khrn\" (UniqueName: \"kubernetes.io/projected/a515b487-9238-4378-82b6-f190dfc360f7-kube-api-access-9khrn\") pod \"ceilometer-0\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.464742 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.476995 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f69bb6594-gttr7" event={"ID":"2dd17bcf-97c4-47f0-86a5-ee28f3a480b6","Type":"ContainerDied","Data":"e845e13b0511fc25d7067f522dddb3f543188107f409fd8c04fc0431e7df2e4b"} Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.477243 4629 scope.go:117] "RemoveContainer" containerID="003b33ac3be999e1aff0c0d9b934346a512a87d1193f1f74b046e59c483c7dda" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.477438 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f69bb6594-gttr7" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.485162 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.485892 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6","Type":"ContainerDied","Data":"854e683489d5a8877b8b72138903fd0037e313cb84512fe25db1b50f24bc22da"} Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.591912 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.595948 4629 scope.go:117] "RemoveContainer" containerID="7d8aee31d5cc0d7710661f70f4cd16fd07e6264d02b55c6e9e30518f043c3bac" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.600447 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.620494 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-f69bb6594-gttr7"] Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.642709 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.662539 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.666817 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.667161 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.668155 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.671178 4629 scope.go:117] "RemoveContainer" containerID="71dd9720317ead0c98a604d8a181df5047d2c718d73a9431208fa124c8e6ccbb" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.688762 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-f69bb6594-gttr7"] Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.722779 4629 scope.go:117] "RemoveContainer" containerID="cb70e2d6df831b4742034249b5c9ed2e0334f2d83b5a83f79954fac414eef11c" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.736397 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.750499 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.750589 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.750727 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdf92\" (UniqueName: \"kubernetes.io/projected/be28a394-a248-4f19-83d0-b7408648fa5d-kube-api-access-xdf92\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.750787 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-config-data\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.750822 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-scripts\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.750900 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-config-data-custom\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.751066 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.751294 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be28a394-a248-4f19-83d0-b7408648fa5d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.751327 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be28a394-a248-4f19-83d0-b7408648fa5d-logs\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.853180 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.853239 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.853316 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdf92\" (UniqueName: \"kubernetes.io/projected/be28a394-a248-4f19-83d0-b7408648fa5d-kube-api-access-xdf92\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.853342 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-config-data\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.853369 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-scripts\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.853398 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-config-data-custom\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.853438 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.857305 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be28a394-a248-4f19-83d0-b7408648fa5d-logs\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.857346 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be28a394-a248-4f19-83d0-b7408648fa5d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.857507 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be28a394-a248-4f19-83d0-b7408648fa5d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.857921 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be28a394-a248-4f19-83d0-b7408648fa5d-logs\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.865325 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.866491 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-config-data-custom\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.866690 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-config-data\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.867337 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-scripts\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.883574 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdf92\" (UniqueName: \"kubernetes.io/projected/be28a394-a248-4f19-83d0-b7408648fa5d-kube-api-access-xdf92\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.899403 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:04 crc kubenswrapper[4629]: I1211 08:59:04.908020 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be28a394-a248-4f19-83d0-b7408648fa5d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"be28a394-a248-4f19-83d0-b7408648fa5d\") " pod="openstack/cinder-api-0" Dec 11 08:59:05 crc kubenswrapper[4629]: I1211 08:59:05.021205 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 08:59:05 crc kubenswrapper[4629]: I1211 08:59:05.110465 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:59:05 crc kubenswrapper[4629]: I1211 08:59:05.329538 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:59:05 crc kubenswrapper[4629]: I1211 08:59:05.552087 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a515b487-9238-4378-82b6-f190dfc360f7","Type":"ContainerStarted","Data":"11e81ef7b32b725459ab18ec170005b399bade1c601e939a6617e613f0a42e2c"} Dec 11 08:59:05 crc kubenswrapper[4629]: I1211 08:59:05.565071 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be28a394-a248-4f19-83d0-b7408648fa5d","Type":"ContainerStarted","Data":"21477719f4363e199b2181700034a92c2280cd942250bcf1630f53f688423233"} Dec 11 08:59:06 crc kubenswrapper[4629]: I1211 08:59:06.209718 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dd17bcf-97c4-47f0-86a5-ee28f3a480b6" path="/var/lib/kubelet/pods/2dd17bcf-97c4-47f0-86a5-ee28f3a480b6/volumes" Dec 11 08:59:06 crc kubenswrapper[4629]: I1211 08:59:06.210778 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6" path="/var/lib/kubelet/pods/ded3f4b5-e297-4ed3-8fcb-9aba9eccb9c6/volumes" Dec 11 08:59:06 crc kubenswrapper[4629]: I1211 08:59:06.589666 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a515b487-9238-4378-82b6-f190dfc360f7","Type":"ContainerStarted","Data":"37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec"} Dec 11 08:59:06 crc kubenswrapper[4629]: I1211 08:59:06.594706 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be28a394-a248-4f19-83d0-b7408648fa5d","Type":"ContainerStarted","Data":"21d86fd0b28e6210d964b2625503893b498868e7906f09f6a8af25abce8cc0b6"} Dec 11 08:59:07 crc kubenswrapper[4629]: I1211 08:59:07.605925 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a515b487-9238-4378-82b6-f190dfc360f7","Type":"ContainerStarted","Data":"ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6"} Dec 11 08:59:07 crc kubenswrapper[4629]: I1211 08:59:07.610715 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be28a394-a248-4f19-83d0-b7408648fa5d","Type":"ContainerStarted","Data":"0c5568bfa028fabef0aa9ca4be03b671897cf8a5db717284c7edce6b20265d53"} Dec 11 08:59:07 crc kubenswrapper[4629]: I1211 08:59:07.610835 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 11 08:59:08 crc kubenswrapper[4629]: I1211 08:59:08.621752 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a515b487-9238-4378-82b6-f190dfc360f7","Type":"ContainerStarted","Data":"5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086"} Dec 11 08:59:08 crc kubenswrapper[4629]: I1211 08:59:08.690468 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.690444499 podStartE2EDuration="4.690444499s" podCreationTimestamp="2025-12-11 08:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:59:07.644939895 +0000 UTC m=+1335.710355503" watchObservedRunningTime="2025-12-11 08:59:08.690444499 +0000 UTC m=+1336.755860127" Dec 11 08:59:08 crc kubenswrapper[4629]: I1211 08:59:08.694723 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.643247 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a515b487-9238-4378-82b6-f190dfc360f7","Type":"ContainerStarted","Data":"34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c"} Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.643625 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.643541 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="ceilometer-central-agent" containerID="cri-o://37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec" gracePeriod=30 Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.643728 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="proxy-httpd" containerID="cri-o://34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c" gracePeriod=30 Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.643817 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="ceilometer-notification-agent" containerID="cri-o://ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6" gracePeriod=30 Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.643896 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="sg-core" containerID="cri-o://5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086" gracePeriod=30 Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.678676 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.5213141220000002 podStartE2EDuration="6.678649921s" podCreationTimestamp="2025-12-11 08:59:04 +0000 UTC" firstStartedPulling="2025-12-11 08:59:05.133052512 +0000 UTC m=+1333.198468120" lastFinishedPulling="2025-12-11 08:59:10.290388311 +0000 UTC m=+1338.355803919" observedRunningTime="2025-12-11 08:59:10.675893314 +0000 UTC m=+1338.741308942" watchObservedRunningTime="2025-12-11 08:59:10.678649921 +0000 UTC m=+1338.744065529" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.719205 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-frlzd"] Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.724775 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-frlzd" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.737347 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-frlzd"] Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.770756 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhqxg\" (UniqueName: \"kubernetes.io/projected/40733a88-03b2-45f0-b982-d6a00fefc036-kube-api-access-jhqxg\") pod \"nova-api-db-create-frlzd\" (UID: \"40733a88-03b2-45f0-b982-d6a00fefc036\") " pod="openstack/nova-api-db-create-frlzd" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.770811 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40733a88-03b2-45f0-b982-d6a00fefc036-operator-scripts\") pod \"nova-api-db-create-frlzd\" (UID: \"40733a88-03b2-45f0-b982-d6a00fefc036\") " pod="openstack/nova-api-db-create-frlzd" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.817719 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-2s8wp"] Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.819959 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2s8wp" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.841748 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-2s8wp"] Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.872402 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmzrm\" (UniqueName: \"kubernetes.io/projected/8f784d8f-ce06-4666-8786-02f327054c68-kube-api-access-fmzrm\") pod \"nova-cell0-db-create-2s8wp\" (UID: \"8f784d8f-ce06-4666-8786-02f327054c68\") " pod="openstack/nova-cell0-db-create-2s8wp" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.872635 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhqxg\" (UniqueName: \"kubernetes.io/projected/40733a88-03b2-45f0-b982-d6a00fefc036-kube-api-access-jhqxg\") pod \"nova-api-db-create-frlzd\" (UID: \"40733a88-03b2-45f0-b982-d6a00fefc036\") " pod="openstack/nova-api-db-create-frlzd" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.872686 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40733a88-03b2-45f0-b982-d6a00fefc036-operator-scripts\") pod \"nova-api-db-create-frlzd\" (UID: \"40733a88-03b2-45f0-b982-d6a00fefc036\") " pod="openstack/nova-api-db-create-frlzd" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.872785 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f784d8f-ce06-4666-8786-02f327054c68-operator-scripts\") pod \"nova-cell0-db-create-2s8wp\" (UID: \"8f784d8f-ce06-4666-8786-02f327054c68\") " pod="openstack/nova-cell0-db-create-2s8wp" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.873482 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40733a88-03b2-45f0-b982-d6a00fefc036-operator-scripts\") pod \"nova-api-db-create-frlzd\" (UID: \"40733a88-03b2-45f0-b982-d6a00fefc036\") " pod="openstack/nova-api-db-create-frlzd" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.897474 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhqxg\" (UniqueName: \"kubernetes.io/projected/40733a88-03b2-45f0-b982-d6a00fefc036-kube-api-access-jhqxg\") pod \"nova-api-db-create-frlzd\" (UID: \"40733a88-03b2-45f0-b982-d6a00fefc036\") " pod="openstack/nova-api-db-create-frlzd" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.932651 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-6448-account-create-update-jgskv"] Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.933732 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6448-account-create-update-jgskv" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.935569 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.948794 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6448-account-create-update-jgskv"] Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.979327 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5qzw\" (UniqueName: \"kubernetes.io/projected/8f738746-49e3-4f84-a85d-124100f4f9de-kube-api-access-h5qzw\") pod \"nova-api-6448-account-create-update-jgskv\" (UID: \"8f738746-49e3-4f84-a85d-124100f4f9de\") " pod="openstack/nova-api-6448-account-create-update-jgskv" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.979430 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmzrm\" (UniqueName: \"kubernetes.io/projected/8f784d8f-ce06-4666-8786-02f327054c68-kube-api-access-fmzrm\") pod \"nova-cell0-db-create-2s8wp\" (UID: \"8f784d8f-ce06-4666-8786-02f327054c68\") " pod="openstack/nova-cell0-db-create-2s8wp" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.979491 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f738746-49e3-4f84-a85d-124100f4f9de-operator-scripts\") pod \"nova-api-6448-account-create-update-jgskv\" (UID: \"8f738746-49e3-4f84-a85d-124100f4f9de\") " pod="openstack/nova-api-6448-account-create-update-jgskv" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.979602 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f784d8f-ce06-4666-8786-02f327054c68-operator-scripts\") pod \"nova-cell0-db-create-2s8wp\" (UID: \"8f784d8f-ce06-4666-8786-02f327054c68\") " pod="openstack/nova-cell0-db-create-2s8wp" Dec 11 08:59:10 crc kubenswrapper[4629]: I1211 08:59:10.980442 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f784d8f-ce06-4666-8786-02f327054c68-operator-scripts\") pod \"nova-cell0-db-create-2s8wp\" (UID: \"8f784d8f-ce06-4666-8786-02f327054c68\") " pod="openstack/nova-cell0-db-create-2s8wp" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.024167 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmzrm\" (UniqueName: \"kubernetes.io/projected/8f784d8f-ce06-4666-8786-02f327054c68-kube-api-access-fmzrm\") pod \"nova-cell0-db-create-2s8wp\" (UID: \"8f784d8f-ce06-4666-8786-02f327054c68\") " pod="openstack/nova-cell0-db-create-2s8wp" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.040608 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-frlzd" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.050358 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-lvfl4"] Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.054143 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-lvfl4" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.060496 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-lvfl4"] Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.085074 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f738746-49e3-4f84-a85d-124100f4f9de-operator-scripts\") pod \"nova-api-6448-account-create-update-jgskv\" (UID: \"8f738746-49e3-4f84-a85d-124100f4f9de\") " pod="openstack/nova-api-6448-account-create-update-jgskv" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.085238 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5qzw\" (UniqueName: \"kubernetes.io/projected/8f738746-49e3-4f84-a85d-124100f4f9de-kube-api-access-h5qzw\") pod \"nova-api-6448-account-create-update-jgskv\" (UID: \"8f738746-49e3-4f84-a85d-124100f4f9de\") " pod="openstack/nova-api-6448-account-create-update-jgskv" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.086208 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f738746-49e3-4f84-a85d-124100f4f9de-operator-scripts\") pod \"nova-api-6448-account-create-update-jgskv\" (UID: \"8f738746-49e3-4f84-a85d-124100f4f9de\") " pod="openstack/nova-api-6448-account-create-update-jgskv" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.123014 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5qzw\" (UniqueName: \"kubernetes.io/projected/8f738746-49e3-4f84-a85d-124100f4f9de-kube-api-access-h5qzw\") pod \"nova-api-6448-account-create-update-jgskv\" (UID: \"8f738746-49e3-4f84-a85d-124100f4f9de\") " pod="openstack/nova-api-6448-account-create-update-jgskv" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.137754 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-8a9c-account-create-update-f8jwt"] Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.138985 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.142152 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2s8wp" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.142217 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.153685 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8a9c-account-create-update-f8jwt"] Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.188766 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcdb2ea1-050a-4541-af4d-db3f2166f578-operator-scripts\") pod \"nova-cell0-8a9c-account-create-update-f8jwt\" (UID: \"bcdb2ea1-050a-4541-af4d-db3f2166f578\") " pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.188874 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373ffecc-c412-4c1c-9451-6c19a39cd8b3-operator-scripts\") pod \"nova-cell1-db-create-lvfl4\" (UID: \"373ffecc-c412-4c1c-9451-6c19a39cd8b3\") " pod="openstack/nova-cell1-db-create-lvfl4" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.188932 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pln8\" (UniqueName: \"kubernetes.io/projected/bcdb2ea1-050a-4541-af4d-db3f2166f578-kube-api-access-4pln8\") pod \"nova-cell0-8a9c-account-create-update-f8jwt\" (UID: \"bcdb2ea1-050a-4541-af4d-db3f2166f578\") " pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.188963 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7vf2\" (UniqueName: \"kubernetes.io/projected/373ffecc-c412-4c1c-9451-6c19a39cd8b3-kube-api-access-g7vf2\") pod \"nova-cell1-db-create-lvfl4\" (UID: \"373ffecc-c412-4c1c-9451-6c19a39cd8b3\") " pod="openstack/nova-cell1-db-create-lvfl4" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.259592 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6448-account-create-update-jgskv" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.290830 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pln8\" (UniqueName: \"kubernetes.io/projected/bcdb2ea1-050a-4541-af4d-db3f2166f578-kube-api-access-4pln8\") pod \"nova-cell0-8a9c-account-create-update-f8jwt\" (UID: \"bcdb2ea1-050a-4541-af4d-db3f2166f578\") " pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.290902 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7vf2\" (UniqueName: \"kubernetes.io/projected/373ffecc-c412-4c1c-9451-6c19a39cd8b3-kube-api-access-g7vf2\") pod \"nova-cell1-db-create-lvfl4\" (UID: \"373ffecc-c412-4c1c-9451-6c19a39cd8b3\") " pod="openstack/nova-cell1-db-create-lvfl4" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.290975 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcdb2ea1-050a-4541-af4d-db3f2166f578-operator-scripts\") pod \"nova-cell0-8a9c-account-create-update-f8jwt\" (UID: \"bcdb2ea1-050a-4541-af4d-db3f2166f578\") " pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.291077 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373ffecc-c412-4c1c-9451-6c19a39cd8b3-operator-scripts\") pod \"nova-cell1-db-create-lvfl4\" (UID: \"373ffecc-c412-4c1c-9451-6c19a39cd8b3\") " pod="openstack/nova-cell1-db-create-lvfl4" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.292069 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373ffecc-c412-4c1c-9451-6c19a39cd8b3-operator-scripts\") pod \"nova-cell1-db-create-lvfl4\" (UID: \"373ffecc-c412-4c1c-9451-6c19a39cd8b3\") " pod="openstack/nova-cell1-db-create-lvfl4" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.293208 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcdb2ea1-050a-4541-af4d-db3f2166f578-operator-scripts\") pod \"nova-cell0-8a9c-account-create-update-f8jwt\" (UID: \"bcdb2ea1-050a-4541-af4d-db3f2166f578\") " pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.318977 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pln8\" (UniqueName: \"kubernetes.io/projected/bcdb2ea1-050a-4541-af4d-db3f2166f578-kube-api-access-4pln8\") pod \"nova-cell0-8a9c-account-create-update-f8jwt\" (UID: \"bcdb2ea1-050a-4541-af4d-db3f2166f578\") " pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.349389 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-7a56-account-create-update-kxj5q"] Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.350472 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.362625 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7vf2\" (UniqueName: \"kubernetes.io/projected/373ffecc-c412-4c1c-9451-6c19a39cd8b3-kube-api-access-g7vf2\") pod \"nova-cell1-db-create-lvfl4\" (UID: \"373ffecc-c412-4c1c-9451-6c19a39cd8b3\") " pod="openstack/nova-cell1-db-create-lvfl4" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.363968 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.384997 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-lvfl4" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.399928 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kffww\" (UniqueName: \"kubernetes.io/projected/6fe1ea45-117b-45a7-bae8-2ef0f78660f2-kube-api-access-kffww\") pod \"nova-cell1-7a56-account-create-update-kxj5q\" (UID: \"6fe1ea45-117b-45a7-bae8-2ef0f78660f2\") " pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.400011 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fe1ea45-117b-45a7-bae8-2ef0f78660f2-operator-scripts\") pod \"nova-cell1-7a56-account-create-update-kxj5q\" (UID: \"6fe1ea45-117b-45a7-bae8-2ef0f78660f2\") " pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.418286 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7a56-account-create-update-kxj5q"] Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.462124 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.501942 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fe1ea45-117b-45a7-bae8-2ef0f78660f2-operator-scripts\") pod \"nova-cell1-7a56-account-create-update-kxj5q\" (UID: \"6fe1ea45-117b-45a7-bae8-2ef0f78660f2\") " pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.502404 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kffww\" (UniqueName: \"kubernetes.io/projected/6fe1ea45-117b-45a7-bae8-2ef0f78660f2-kube-api-access-kffww\") pod \"nova-cell1-7a56-account-create-update-kxj5q\" (UID: \"6fe1ea45-117b-45a7-bae8-2ef0f78660f2\") " pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.503260 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fe1ea45-117b-45a7-bae8-2ef0f78660f2-operator-scripts\") pod \"nova-cell1-7a56-account-create-update-kxj5q\" (UID: \"6fe1ea45-117b-45a7-bae8-2ef0f78660f2\") " pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.524712 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kffww\" (UniqueName: \"kubernetes.io/projected/6fe1ea45-117b-45a7-bae8-2ef0f78660f2-kube-api-access-kffww\") pod \"nova-cell1-7a56-account-create-update-kxj5q\" (UID: \"6fe1ea45-117b-45a7-bae8-2ef0f78660f2\") " pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.669537 4629 generic.go:334] "Generic (PLEG): container finished" podID="a515b487-9238-4378-82b6-f190dfc360f7" containerID="5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086" exitCode=2 Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.672059 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a515b487-9238-4378-82b6-f190dfc360f7","Type":"ContainerDied","Data":"5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086"} Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.720900 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-frlzd"] Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.811192 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" Dec 11 08:59:11 crc kubenswrapper[4629]: I1211 08:59:11.992785 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-2s8wp"] Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.077495 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8a9c-account-create-update-f8jwt"] Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.111791 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6448-account-create-update-jgskv"] Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.238356 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-lvfl4"] Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.565593 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7a56-account-create-update-kxj5q"] Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.688153 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" event={"ID":"bcdb2ea1-050a-4541-af4d-db3f2166f578","Type":"ContainerStarted","Data":"02d45d32851e1fba92a031ac9d4238fd31a4597d1ca86bc1db68b2430461d574"} Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.689040 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-lvfl4" event={"ID":"373ffecc-c412-4c1c-9451-6c19a39cd8b3","Type":"ContainerStarted","Data":"afb30624b61237556d5e9ed5b041bd9d3ffaaed58966d178adf8eda5100d51eb"} Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.690043 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2s8wp" event={"ID":"8f784d8f-ce06-4666-8786-02f327054c68","Type":"ContainerStarted","Data":"3d6b02d32cae382d0020b0b30f12e0fddbfabd7aaf0cebb386145524866a6ad7"} Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.690069 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2s8wp" event={"ID":"8f784d8f-ce06-4666-8786-02f327054c68","Type":"ContainerStarted","Data":"9223a9a5cfed16a15f205a898153bea0d17f27561562c4ce2e2ca1b6becc4d4c"} Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.692316 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6448-account-create-update-jgskv" event={"ID":"8f738746-49e3-4f84-a85d-124100f4f9de","Type":"ContainerStarted","Data":"b54680def0b361ac60532ee44533facacf12134dff50230afdb5ce81600409f5"} Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.703033 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-frlzd" event={"ID":"40733a88-03b2-45f0-b982-d6a00fefc036","Type":"ContainerStarted","Data":"fd100bb683f86cee57c19258890d16d1c0566b0c2ac3edc7b27b138f76273aa3"} Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.703090 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-frlzd" event={"ID":"40733a88-03b2-45f0-b982-d6a00fefc036","Type":"ContainerStarted","Data":"6daad46574fd2029994f234f7fe66e7e67f982e56c1dd1796a93afc6ed75ddd3"} Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.722144 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" event={"ID":"6fe1ea45-117b-45a7-bae8-2ef0f78660f2","Type":"ContainerStarted","Data":"e632a0339eebe1558335a79b31696f7d29c278308b48269c3b9380ea8f6b3cfe"} Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.744675 4629 generic.go:334] "Generic (PLEG): container finished" podID="a515b487-9238-4378-82b6-f190dfc360f7" containerID="ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6" exitCode=0 Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.748137 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a515b487-9238-4378-82b6-f190dfc360f7","Type":"ContainerDied","Data":"ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6"} Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.762292 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-frlzd" podStartSLOduration=2.7622760189999997 podStartE2EDuration="2.762276019s" podCreationTimestamp="2025-12-11 08:59:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:59:12.759718729 +0000 UTC m=+1340.825134337" watchObservedRunningTime="2025-12-11 08:59:12.762276019 +0000 UTC m=+1340.827691627" Dec 11 08:59:12 crc kubenswrapper[4629]: I1211 08:59:12.763075 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-2s8wp" podStartSLOduration=2.763069574 podStartE2EDuration="2.763069574s" podCreationTimestamp="2025-12-11 08:59:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:59:12.739259236 +0000 UTC m=+1340.804674844" watchObservedRunningTime="2025-12-11 08:59:12.763069574 +0000 UTC m=+1340.828485182" Dec 11 08:59:13 crc kubenswrapper[4629]: I1211 08:59:13.754697 4629 generic.go:334] "Generic (PLEG): container finished" podID="bcdb2ea1-050a-4541-af4d-db3f2166f578" containerID="828f2f621c907f59d2698842ac2e387d0a2be413afe971e677dde5f123ae7eca" exitCode=0 Dec 11 08:59:13 crc kubenswrapper[4629]: I1211 08:59:13.755122 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" event={"ID":"bcdb2ea1-050a-4541-af4d-db3f2166f578","Type":"ContainerDied","Data":"828f2f621c907f59d2698842ac2e387d0a2be413afe971e677dde5f123ae7eca"} Dec 11 08:59:13 crc kubenswrapper[4629]: I1211 08:59:13.758306 4629 generic.go:334] "Generic (PLEG): container finished" podID="373ffecc-c412-4c1c-9451-6c19a39cd8b3" containerID="8698a7dcb6d1bf64f0e490c80eb1f24ee35ba2acee78dc41795884c93867b8b1" exitCode=0 Dec 11 08:59:13 crc kubenswrapper[4629]: I1211 08:59:13.758430 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-lvfl4" event={"ID":"373ffecc-c412-4c1c-9451-6c19a39cd8b3","Type":"ContainerDied","Data":"8698a7dcb6d1bf64f0e490c80eb1f24ee35ba2acee78dc41795884c93867b8b1"} Dec 11 08:59:13 crc kubenswrapper[4629]: I1211 08:59:13.759901 4629 generic.go:334] "Generic (PLEG): container finished" podID="8f784d8f-ce06-4666-8786-02f327054c68" containerID="3d6b02d32cae382d0020b0b30f12e0fddbfabd7aaf0cebb386145524866a6ad7" exitCode=0 Dec 11 08:59:13 crc kubenswrapper[4629]: I1211 08:59:13.759970 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2s8wp" event={"ID":"8f784d8f-ce06-4666-8786-02f327054c68","Type":"ContainerDied","Data":"3d6b02d32cae382d0020b0b30f12e0fddbfabd7aaf0cebb386145524866a6ad7"} Dec 11 08:59:13 crc kubenswrapper[4629]: I1211 08:59:13.761496 4629 generic.go:334] "Generic (PLEG): container finished" podID="8f738746-49e3-4f84-a85d-124100f4f9de" containerID="fe76614364efaac6558506d4ef88813004df0129c350fd46648ae5610a1205fe" exitCode=0 Dec 11 08:59:13 crc kubenswrapper[4629]: I1211 08:59:13.761563 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6448-account-create-update-jgskv" event={"ID":"8f738746-49e3-4f84-a85d-124100f4f9de","Type":"ContainerDied","Data":"fe76614364efaac6558506d4ef88813004df0129c350fd46648ae5610a1205fe"} Dec 11 08:59:13 crc kubenswrapper[4629]: I1211 08:59:13.763350 4629 generic.go:334] "Generic (PLEG): container finished" podID="40733a88-03b2-45f0-b982-d6a00fefc036" containerID="fd100bb683f86cee57c19258890d16d1c0566b0c2ac3edc7b27b138f76273aa3" exitCode=0 Dec 11 08:59:13 crc kubenswrapper[4629]: I1211 08:59:13.763425 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-frlzd" event={"ID":"40733a88-03b2-45f0-b982-d6a00fefc036","Type":"ContainerDied","Data":"fd100bb683f86cee57c19258890d16d1c0566b0c2ac3edc7b27b138f76273aa3"} Dec 11 08:59:13 crc kubenswrapper[4629]: I1211 08:59:13.767932 4629 generic.go:334] "Generic (PLEG): container finished" podID="6fe1ea45-117b-45a7-bae8-2ef0f78660f2" containerID="6cecbef414779753a6e47aa24ccfed626aecffadf7dfa358b2f274ffa2d77a10" exitCode=0 Dec 11 08:59:13 crc kubenswrapper[4629]: I1211 08:59:13.767985 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" event={"ID":"6fe1ea45-117b-45a7-bae8-2ef0f78660f2","Type":"ContainerDied","Data":"6cecbef414779753a6e47aa24ccfed626aecffadf7dfa358b2f274ffa2d77a10"} Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.280586 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.394804 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pln8\" (UniqueName: \"kubernetes.io/projected/bcdb2ea1-050a-4541-af4d-db3f2166f578-kube-api-access-4pln8\") pod \"bcdb2ea1-050a-4541-af4d-db3f2166f578\" (UID: \"bcdb2ea1-050a-4541-af4d-db3f2166f578\") " Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.394896 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcdb2ea1-050a-4541-af4d-db3f2166f578-operator-scripts\") pod \"bcdb2ea1-050a-4541-af4d-db3f2166f578\" (UID: \"bcdb2ea1-050a-4541-af4d-db3f2166f578\") " Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.396108 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcdb2ea1-050a-4541-af4d-db3f2166f578-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bcdb2ea1-050a-4541-af4d-db3f2166f578" (UID: "bcdb2ea1-050a-4541-af4d-db3f2166f578"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.416159 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcdb2ea1-050a-4541-af4d-db3f2166f578-kube-api-access-4pln8" (OuterVolumeSpecName: "kube-api-access-4pln8") pod "bcdb2ea1-050a-4541-af4d-db3f2166f578" (UID: "bcdb2ea1-050a-4541-af4d-db3f2166f578"). InnerVolumeSpecName "kube-api-access-4pln8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.497272 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pln8\" (UniqueName: \"kubernetes.io/projected/bcdb2ea1-050a-4541-af4d-db3f2166f578-kube-api-access-4pln8\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.497335 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcdb2ea1-050a-4541-af4d-db3f2166f578-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.549356 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-frlzd" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.568598 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-lvfl4" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.571093 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6448-account-create-update-jgskv" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.582107 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2s8wp" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.590443 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.705565 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhqxg\" (UniqueName: \"kubernetes.io/projected/40733a88-03b2-45f0-b982-d6a00fefc036-kube-api-access-jhqxg\") pod \"40733a88-03b2-45f0-b982-d6a00fefc036\" (UID: \"40733a88-03b2-45f0-b982-d6a00fefc036\") " Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.705676 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40733a88-03b2-45f0-b982-d6a00fefc036-operator-scripts\") pod \"40733a88-03b2-45f0-b982-d6a00fefc036\" (UID: \"40733a88-03b2-45f0-b982-d6a00fefc036\") " Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.705718 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f738746-49e3-4f84-a85d-124100f4f9de-operator-scripts\") pod \"8f738746-49e3-4f84-a85d-124100f4f9de\" (UID: \"8f738746-49e3-4f84-a85d-124100f4f9de\") " Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.705772 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fe1ea45-117b-45a7-bae8-2ef0f78660f2-operator-scripts\") pod \"6fe1ea45-117b-45a7-bae8-2ef0f78660f2\" (UID: \"6fe1ea45-117b-45a7-bae8-2ef0f78660f2\") " Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.705799 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f784d8f-ce06-4666-8786-02f327054c68-operator-scripts\") pod \"8f784d8f-ce06-4666-8786-02f327054c68\" (UID: \"8f784d8f-ce06-4666-8786-02f327054c68\") " Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.705925 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7vf2\" (UniqueName: \"kubernetes.io/projected/373ffecc-c412-4c1c-9451-6c19a39cd8b3-kube-api-access-g7vf2\") pod \"373ffecc-c412-4c1c-9451-6c19a39cd8b3\" (UID: \"373ffecc-c412-4c1c-9451-6c19a39cd8b3\") " Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.705971 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kffww\" (UniqueName: \"kubernetes.io/projected/6fe1ea45-117b-45a7-bae8-2ef0f78660f2-kube-api-access-kffww\") pod \"6fe1ea45-117b-45a7-bae8-2ef0f78660f2\" (UID: \"6fe1ea45-117b-45a7-bae8-2ef0f78660f2\") " Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.706043 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmzrm\" (UniqueName: \"kubernetes.io/projected/8f784d8f-ce06-4666-8786-02f327054c68-kube-api-access-fmzrm\") pod \"8f784d8f-ce06-4666-8786-02f327054c68\" (UID: \"8f784d8f-ce06-4666-8786-02f327054c68\") " Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.706073 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373ffecc-c412-4c1c-9451-6c19a39cd8b3-operator-scripts\") pod \"373ffecc-c412-4c1c-9451-6c19a39cd8b3\" (UID: \"373ffecc-c412-4c1c-9451-6c19a39cd8b3\") " Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.706095 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5qzw\" (UniqueName: \"kubernetes.io/projected/8f738746-49e3-4f84-a85d-124100f4f9de-kube-api-access-h5qzw\") pod \"8f738746-49e3-4f84-a85d-124100f4f9de\" (UID: \"8f738746-49e3-4f84-a85d-124100f4f9de\") " Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.706917 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f784d8f-ce06-4666-8786-02f327054c68-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8f784d8f-ce06-4666-8786-02f327054c68" (UID: "8f784d8f-ce06-4666-8786-02f327054c68"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.708378 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f738746-49e3-4f84-a85d-124100f4f9de-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8f738746-49e3-4f84-a85d-124100f4f9de" (UID: "8f738746-49e3-4f84-a85d-124100f4f9de"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.708692 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40733a88-03b2-45f0-b982-d6a00fefc036-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "40733a88-03b2-45f0-b982-d6a00fefc036" (UID: "40733a88-03b2-45f0-b982-d6a00fefc036"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.709009 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe1ea45-117b-45a7-bae8-2ef0f78660f2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6fe1ea45-117b-45a7-bae8-2ef0f78660f2" (UID: "6fe1ea45-117b-45a7-bae8-2ef0f78660f2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.710946 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40733a88-03b2-45f0-b982-d6a00fefc036-kube-api-access-jhqxg" (OuterVolumeSpecName: "kube-api-access-jhqxg") pod "40733a88-03b2-45f0-b982-d6a00fefc036" (UID: "40733a88-03b2-45f0-b982-d6a00fefc036"). InnerVolumeSpecName "kube-api-access-jhqxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.711501 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/373ffecc-c412-4c1c-9451-6c19a39cd8b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "373ffecc-c412-4c1c-9451-6c19a39cd8b3" (UID: "373ffecc-c412-4c1c-9451-6c19a39cd8b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.718074 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fe1ea45-117b-45a7-bae8-2ef0f78660f2-kube-api-access-kffww" (OuterVolumeSpecName: "kube-api-access-kffww") pod "6fe1ea45-117b-45a7-bae8-2ef0f78660f2" (UID: "6fe1ea45-117b-45a7-bae8-2ef0f78660f2"). InnerVolumeSpecName "kube-api-access-kffww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.718206 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f784d8f-ce06-4666-8786-02f327054c68-kube-api-access-fmzrm" (OuterVolumeSpecName: "kube-api-access-fmzrm") pod "8f784d8f-ce06-4666-8786-02f327054c68" (UID: "8f784d8f-ce06-4666-8786-02f327054c68"). InnerVolumeSpecName "kube-api-access-fmzrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.718264 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f738746-49e3-4f84-a85d-124100f4f9de-kube-api-access-h5qzw" (OuterVolumeSpecName: "kube-api-access-h5qzw") pod "8f738746-49e3-4f84-a85d-124100f4f9de" (UID: "8f738746-49e3-4f84-a85d-124100f4f9de"). InnerVolumeSpecName "kube-api-access-h5qzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.726077 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/373ffecc-c412-4c1c-9451-6c19a39cd8b3-kube-api-access-g7vf2" (OuterVolumeSpecName: "kube-api-access-g7vf2") pod "373ffecc-c412-4c1c-9451-6c19a39cd8b3" (UID: "373ffecc-c412-4c1c-9451-6c19a39cd8b3"). InnerVolumeSpecName "kube-api-access-g7vf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.786910 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-lvfl4" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.786922 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-lvfl4" event={"ID":"373ffecc-c412-4c1c-9451-6c19a39cd8b3","Type":"ContainerDied","Data":"afb30624b61237556d5e9ed5b041bd9d3ffaaed58966d178adf8eda5100d51eb"} Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.786964 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afb30624b61237556d5e9ed5b041bd9d3ffaaed58966d178adf8eda5100d51eb" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.788442 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2s8wp" event={"ID":"8f784d8f-ce06-4666-8786-02f327054c68","Type":"ContainerDied","Data":"9223a9a5cfed16a15f205a898153bea0d17f27561562c4ce2e2ca1b6becc4d4c"} Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.788466 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9223a9a5cfed16a15f205a898153bea0d17f27561562c4ce2e2ca1b6becc4d4c" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.788530 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2s8wp" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.793570 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6448-account-create-update-jgskv" event={"ID":"8f738746-49e3-4f84-a85d-124100f4f9de","Type":"ContainerDied","Data":"b54680def0b361ac60532ee44533facacf12134dff50230afdb5ce81600409f5"} Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.793612 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b54680def0b361ac60532ee44533facacf12134dff50230afdb5ce81600409f5" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.793658 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6448-account-create-update-jgskv" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.797744 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-frlzd" event={"ID":"40733a88-03b2-45f0-b982-d6a00fefc036","Type":"ContainerDied","Data":"6daad46574fd2029994f234f7fe66e7e67f982e56c1dd1796a93afc6ed75ddd3"} Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.797789 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6daad46574fd2029994f234f7fe66e7e67f982e56c1dd1796a93afc6ed75ddd3" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.797866 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-frlzd" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.799382 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" event={"ID":"6fe1ea45-117b-45a7-bae8-2ef0f78660f2","Type":"ContainerDied","Data":"e632a0339eebe1558335a79b31696f7d29c278308b48269c3b9380ea8f6b3cfe"} Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.799417 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e632a0339eebe1558335a79b31696f7d29c278308b48269c3b9380ea8f6b3cfe" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.799473 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7a56-account-create-update-kxj5q" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.805238 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" event={"ID":"bcdb2ea1-050a-4541-af4d-db3f2166f578","Type":"ContainerDied","Data":"02d45d32851e1fba92a031ac9d4238fd31a4597d1ca86bc1db68b2430461d574"} Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.805281 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02d45d32851e1fba92a031ac9d4238fd31a4597d1ca86bc1db68b2430461d574" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.805304 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8a9c-account-create-update-f8jwt" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.808341 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fe1ea45-117b-45a7-bae8-2ef0f78660f2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.808370 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f784d8f-ce06-4666-8786-02f327054c68-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.808383 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7vf2\" (UniqueName: \"kubernetes.io/projected/373ffecc-c412-4c1c-9451-6c19a39cd8b3-kube-api-access-g7vf2\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.808392 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kffww\" (UniqueName: \"kubernetes.io/projected/6fe1ea45-117b-45a7-bae8-2ef0f78660f2-kube-api-access-kffww\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.808400 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmzrm\" (UniqueName: \"kubernetes.io/projected/8f784d8f-ce06-4666-8786-02f327054c68-kube-api-access-fmzrm\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.808408 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373ffecc-c412-4c1c-9451-6c19a39cd8b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.808417 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5qzw\" (UniqueName: \"kubernetes.io/projected/8f738746-49e3-4f84-a85d-124100f4f9de-kube-api-access-h5qzw\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.808427 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhqxg\" (UniqueName: \"kubernetes.io/projected/40733a88-03b2-45f0-b982-d6a00fefc036-kube-api-access-jhqxg\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.808435 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40733a88-03b2-45f0-b982-d6a00fefc036-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:15 crc kubenswrapper[4629]: I1211 08:59:15.808443 4629 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f738746-49e3-4f84-a85d-124100f4f9de-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:16 crc kubenswrapper[4629]: I1211 08:59:16.816181 4629 generic.go:334] "Generic (PLEG): container finished" podID="a515b487-9238-4378-82b6-f190dfc360f7" containerID="37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec" exitCode=0 Dec 11 08:59:16 crc kubenswrapper[4629]: I1211 08:59:16.816259 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a515b487-9238-4378-82b6-f190dfc360f7","Type":"ContainerDied","Data":"37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec"} Dec 11 08:59:17 crc kubenswrapper[4629]: I1211 08:59:17.470830 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 11 08:59:20 crc kubenswrapper[4629]: I1211 08:59:20.486264 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:59:20 crc kubenswrapper[4629]: I1211 08:59:20.486626 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.440806 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g6gr5"] Dec 11 08:59:21 crc kubenswrapper[4629]: E1211 08:59:21.442021 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373ffecc-c412-4c1c-9451-6c19a39cd8b3" containerName="mariadb-database-create" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.442119 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="373ffecc-c412-4c1c-9451-6c19a39cd8b3" containerName="mariadb-database-create" Dec 11 08:59:21 crc kubenswrapper[4629]: E1211 08:59:21.442204 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fe1ea45-117b-45a7-bae8-2ef0f78660f2" containerName="mariadb-account-create-update" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.442277 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fe1ea45-117b-45a7-bae8-2ef0f78660f2" containerName="mariadb-account-create-update" Dec 11 08:59:21 crc kubenswrapper[4629]: E1211 08:59:21.442361 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f784d8f-ce06-4666-8786-02f327054c68" containerName="mariadb-database-create" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.442447 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f784d8f-ce06-4666-8786-02f327054c68" containerName="mariadb-database-create" Dec 11 08:59:21 crc kubenswrapper[4629]: E1211 08:59:21.442548 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcdb2ea1-050a-4541-af4d-db3f2166f578" containerName="mariadb-account-create-update" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.442653 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcdb2ea1-050a-4541-af4d-db3f2166f578" containerName="mariadb-account-create-update" Dec 11 08:59:21 crc kubenswrapper[4629]: E1211 08:59:21.442741 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40733a88-03b2-45f0-b982-d6a00fefc036" containerName="mariadb-database-create" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.442813 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="40733a88-03b2-45f0-b982-d6a00fefc036" containerName="mariadb-database-create" Dec 11 08:59:21 crc kubenswrapper[4629]: E1211 08:59:21.442910 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f738746-49e3-4f84-a85d-124100f4f9de" containerName="mariadb-account-create-update" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.442984 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f738746-49e3-4f84-a85d-124100f4f9de" containerName="mariadb-account-create-update" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.443280 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="40733a88-03b2-45f0-b982-d6a00fefc036" containerName="mariadb-database-create" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.443380 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="373ffecc-c412-4c1c-9451-6c19a39cd8b3" containerName="mariadb-database-create" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.443465 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fe1ea45-117b-45a7-bae8-2ef0f78660f2" containerName="mariadb-account-create-update" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.443553 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcdb2ea1-050a-4541-af4d-db3f2166f578" containerName="mariadb-account-create-update" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.443644 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f738746-49e3-4f84-a85d-124100f4f9de" containerName="mariadb-account-create-update" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.443726 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f784d8f-ce06-4666-8786-02f327054c68" containerName="mariadb-database-create" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.444574 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.449922 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.450365 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.452250 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gb65z" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.460631 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g6gr5"] Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.510915 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-scripts\") pod \"nova-cell0-conductor-db-sync-g6gr5\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.511909 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqmjr\" (UniqueName: \"kubernetes.io/projected/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-kube-api-access-zqmjr\") pod \"nova-cell0-conductor-db-sync-g6gr5\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.512112 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-config-data\") pod \"nova-cell0-conductor-db-sync-g6gr5\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.512219 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-g6gr5\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.614048 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqmjr\" (UniqueName: \"kubernetes.io/projected/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-kube-api-access-zqmjr\") pod \"nova-cell0-conductor-db-sync-g6gr5\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.614664 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-config-data\") pod \"nova-cell0-conductor-db-sync-g6gr5\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.615658 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-g6gr5\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.615788 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-scripts\") pod \"nova-cell0-conductor-db-sync-g6gr5\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.622508 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-g6gr5\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.632350 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-scripts\") pod \"nova-cell0-conductor-db-sync-g6gr5\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.633429 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-config-data\") pod \"nova-cell0-conductor-db-sync-g6gr5\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.655838 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqmjr\" (UniqueName: \"kubernetes.io/projected/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-kube-api-access-zqmjr\") pod \"nova-cell0-conductor-db-sync-g6gr5\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:21 crc kubenswrapper[4629]: I1211 08:59:21.766544 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:23 crc kubenswrapper[4629]: W1211 08:59:23.003346 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45f8baeb_687e_4c40_8ea9_05e66a26c9a7.slice/crio-b67187f47d38e912608f0a5e1069575b7341a197607cf00490a3cda2e3b33cd0 WatchSource:0}: Error finding container b67187f47d38e912608f0a5e1069575b7341a197607cf00490a3cda2e3b33cd0: Status 404 returned error can't find the container with id b67187f47d38e912608f0a5e1069575b7341a197607cf00490a3cda2e3b33cd0 Dec 11 08:59:23 crc kubenswrapper[4629]: I1211 08:59:23.005782 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g6gr5"] Dec 11 08:59:23 crc kubenswrapper[4629]: I1211 08:59:23.891868 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-g6gr5" event={"ID":"45f8baeb-687e-4c40-8ea9-05e66a26c9a7","Type":"ContainerStarted","Data":"b67187f47d38e912608f0a5e1069575b7341a197607cf00490a3cda2e3b33cd0"} Dec 11 08:59:33 crc kubenswrapper[4629]: I1211 08:59:33.986747 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-g6gr5" event={"ID":"45f8baeb-687e-4c40-8ea9-05e66a26c9a7","Type":"ContainerStarted","Data":"91ffd72b0961ed31745502586d4d8ddfc25313b7cf1442f3495c7b21f64208c5"} Dec 11 08:59:34 crc kubenswrapper[4629]: I1211 08:59:34.012769 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-g6gr5" podStartSLOduration=3.310269557 podStartE2EDuration="13.012750474s" podCreationTimestamp="2025-12-11 08:59:21 +0000 UTC" firstStartedPulling="2025-12-11 08:59:23.00635089 +0000 UTC m=+1351.071766498" lastFinishedPulling="2025-12-11 08:59:32.708831807 +0000 UTC m=+1360.774247415" observedRunningTime="2025-12-11 08:59:34.012147676 +0000 UTC m=+1362.077563284" watchObservedRunningTime="2025-12-11 08:59:34.012750474 +0000 UTC m=+1362.078166082" Dec 11 08:59:34 crc kubenswrapper[4629]: I1211 08:59:34.470219 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.023933 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.059015 4629 generic.go:334] "Generic (PLEG): container finished" podID="a515b487-9238-4378-82b6-f190dfc360f7" containerID="34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c" exitCode=137 Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.059078 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a515b487-9238-4378-82b6-f190dfc360f7","Type":"ContainerDied","Data":"34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c"} Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.059111 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a515b487-9238-4378-82b6-f190dfc360f7","Type":"ContainerDied","Data":"11e81ef7b32b725459ab18ec170005b399bade1c601e939a6617e613f0a42e2c"} Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.059133 4629 scope.go:117] "RemoveContainer" containerID="34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.059302 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.084472 4629 scope.go:117] "RemoveContainer" containerID="5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.102461 4629 scope.go:117] "RemoveContainer" containerID="ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.121590 4629 scope.go:117] "RemoveContainer" containerID="37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.142839 4629 scope.go:117] "RemoveContainer" containerID="34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c" Dec 11 08:59:41 crc kubenswrapper[4629]: E1211 08:59:41.143349 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c\": container with ID starting with 34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c not found: ID does not exist" containerID="34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.143401 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c"} err="failed to get container status \"34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c\": rpc error: code = NotFound desc = could not find container \"34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c\": container with ID starting with 34311d49bd9656e7c39fd5d52cf0af67a1fb19fa5d2c037d89c3370273e6f54c not found: ID does not exist" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.143437 4629 scope.go:117] "RemoveContainer" containerID="5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086" Dec 11 08:59:41 crc kubenswrapper[4629]: E1211 08:59:41.143799 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086\": container with ID starting with 5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086 not found: ID does not exist" containerID="5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.143830 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086"} err="failed to get container status \"5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086\": rpc error: code = NotFound desc = could not find container \"5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086\": container with ID starting with 5bbfba6fa2d08c1def87d5e64938432ad095b2c9934e697889fa4be6c4f03086 not found: ID does not exist" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.143866 4629 scope.go:117] "RemoveContainer" containerID="ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6" Dec 11 08:59:41 crc kubenswrapper[4629]: E1211 08:59:41.144224 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6\": container with ID starting with ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6 not found: ID does not exist" containerID="ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.144248 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6"} err="failed to get container status \"ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6\": rpc error: code = NotFound desc = could not find container \"ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6\": container with ID starting with ec9933577f78440d64bdce49d6dfd23eae262866f6acd90e3d7d611abf0dded6 not found: ID does not exist" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.144280 4629 scope.go:117] "RemoveContainer" containerID="37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec" Dec 11 08:59:41 crc kubenswrapper[4629]: E1211 08:59:41.144760 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec\": container with ID starting with 37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec not found: ID does not exist" containerID="37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.144820 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec"} err="failed to get container status \"37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec\": rpc error: code = NotFound desc = could not find container \"37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec\": container with ID starting with 37bd3c78ee89f8b8530f0a21d3e94c039503bc681020ab81fc06733452e298ec not found: ID does not exist" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.177493 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a515b487-9238-4378-82b6-f190dfc360f7-log-httpd\") pod \"a515b487-9238-4378-82b6-f190dfc360f7\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.177597 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9khrn\" (UniqueName: \"kubernetes.io/projected/a515b487-9238-4378-82b6-f190dfc360f7-kube-api-access-9khrn\") pod \"a515b487-9238-4378-82b6-f190dfc360f7\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.177672 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-combined-ca-bundle\") pod \"a515b487-9238-4378-82b6-f190dfc360f7\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.177725 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-sg-core-conf-yaml\") pod \"a515b487-9238-4378-82b6-f190dfc360f7\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.177740 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a515b487-9238-4378-82b6-f190dfc360f7-run-httpd\") pod \"a515b487-9238-4378-82b6-f190dfc360f7\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.177803 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-scripts\") pod \"a515b487-9238-4378-82b6-f190dfc360f7\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.177912 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-config-data\") pod \"a515b487-9238-4378-82b6-f190dfc360f7\" (UID: \"a515b487-9238-4378-82b6-f190dfc360f7\") " Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.178649 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a515b487-9238-4378-82b6-f190dfc360f7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a515b487-9238-4378-82b6-f190dfc360f7" (UID: "a515b487-9238-4378-82b6-f190dfc360f7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.179196 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a515b487-9238-4378-82b6-f190dfc360f7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a515b487-9238-4378-82b6-f190dfc360f7" (UID: "a515b487-9238-4378-82b6-f190dfc360f7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.185017 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a515b487-9238-4378-82b6-f190dfc360f7-kube-api-access-9khrn" (OuterVolumeSpecName: "kube-api-access-9khrn") pod "a515b487-9238-4378-82b6-f190dfc360f7" (UID: "a515b487-9238-4378-82b6-f190dfc360f7"). InnerVolumeSpecName "kube-api-access-9khrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.198223 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-scripts" (OuterVolumeSpecName: "scripts") pod "a515b487-9238-4378-82b6-f190dfc360f7" (UID: "a515b487-9238-4378-82b6-f190dfc360f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.213253 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a515b487-9238-4378-82b6-f190dfc360f7" (UID: "a515b487-9238-4378-82b6-f190dfc360f7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.254394 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a515b487-9238-4378-82b6-f190dfc360f7" (UID: "a515b487-9238-4378-82b6-f190dfc360f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.273062 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-config-data" (OuterVolumeSpecName: "config-data") pod "a515b487-9238-4378-82b6-f190dfc360f7" (UID: "a515b487-9238-4378-82b6-f190dfc360f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.279483 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.279603 4629 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a515b487-9238-4378-82b6-f190dfc360f7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.279678 4629 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.279736 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.279789 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a515b487-9238-4378-82b6-f190dfc360f7-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.279858 4629 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a515b487-9238-4378-82b6-f190dfc360f7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.279926 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9khrn\" (UniqueName: \"kubernetes.io/projected/a515b487-9238-4378-82b6-f190dfc360f7-kube-api-access-9khrn\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.396708 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.406207 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.426611 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:59:41 crc kubenswrapper[4629]: E1211 08:59:41.427402 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="ceilometer-central-agent" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.427425 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="ceilometer-central-agent" Dec 11 08:59:41 crc kubenswrapper[4629]: E1211 08:59:41.427441 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="sg-core" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.427447 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="sg-core" Dec 11 08:59:41 crc kubenswrapper[4629]: E1211 08:59:41.427467 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="proxy-httpd" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.427475 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="proxy-httpd" Dec 11 08:59:41 crc kubenswrapper[4629]: E1211 08:59:41.427489 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="ceilometer-notification-agent" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.427495 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="ceilometer-notification-agent" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.427836 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="ceilometer-notification-agent" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.427879 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="ceilometer-central-agent" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.427891 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="proxy-httpd" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.427902 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="a515b487-9238-4378-82b6-f190dfc360f7" containerName="sg-core" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.430489 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.439408 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.439678 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.447753 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.585239 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.585612 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ad9572b-797d-4696-bd0a-2f6c388827ff-run-httpd\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.585803 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ad9572b-797d-4696-bd0a-2f6c388827ff-log-httpd\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.586101 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.586164 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-scripts\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.586321 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz5j2\" (UniqueName: \"kubernetes.io/projected/8ad9572b-797d-4696-bd0a-2f6c388827ff-kube-api-access-nz5j2\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.586342 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-config-data\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.688074 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz5j2\" (UniqueName: \"kubernetes.io/projected/8ad9572b-797d-4696-bd0a-2f6c388827ff-kube-api-access-nz5j2\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.688134 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-config-data\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.688161 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.688204 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ad9572b-797d-4696-bd0a-2f6c388827ff-run-httpd\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.688261 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ad9572b-797d-4696-bd0a-2f6c388827ff-log-httpd\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.688317 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.688346 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-scripts\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.688936 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ad9572b-797d-4696-bd0a-2f6c388827ff-log-httpd\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.689138 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ad9572b-797d-4696-bd0a-2f6c388827ff-run-httpd\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.694021 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.694750 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-scripts\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.695753 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.696663 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-config-data\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.710363 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz5j2\" (UniqueName: \"kubernetes.io/projected/8ad9572b-797d-4696-bd0a-2f6c388827ff-kube-api-access-nz5j2\") pod \"ceilometer-0\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " pod="openstack/ceilometer-0" Dec 11 08:59:41 crc kubenswrapper[4629]: I1211 08:59:41.764971 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:59:42 crc kubenswrapper[4629]: I1211 08:59:42.210437 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a515b487-9238-4378-82b6-f190dfc360f7" path="/var/lib/kubelet/pods/a515b487-9238-4378-82b6-f190dfc360f7/volumes" Dec 11 08:59:42 crc kubenswrapper[4629]: I1211 08:59:42.256959 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:59:42 crc kubenswrapper[4629]: W1211 08:59:42.257959 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ad9572b_797d_4696_bd0a_2f6c388827ff.slice/crio-4caf9f8da6169afa1b33348aad1e7e44acce078db64489f1a9dd11513b4b9088 WatchSource:0}: Error finding container 4caf9f8da6169afa1b33348aad1e7e44acce078db64489f1a9dd11513b4b9088: Status 404 returned error can't find the container with id 4caf9f8da6169afa1b33348aad1e7e44acce078db64489f1a9dd11513b4b9088 Dec 11 08:59:42 crc kubenswrapper[4629]: I1211 08:59:42.260629 4629 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 08:59:43 crc kubenswrapper[4629]: I1211 08:59:43.098635 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ad9572b-797d-4696-bd0a-2f6c388827ff","Type":"ContainerStarted","Data":"65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630"} Dec 11 08:59:43 crc kubenswrapper[4629]: I1211 08:59:43.099021 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ad9572b-797d-4696-bd0a-2f6c388827ff","Type":"ContainerStarted","Data":"4caf9f8da6169afa1b33348aad1e7e44acce078db64489f1a9dd11513b4b9088"} Dec 11 08:59:44 crc kubenswrapper[4629]: I1211 08:59:44.108452 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ad9572b-797d-4696-bd0a-2f6c388827ff","Type":"ContainerStarted","Data":"cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205"} Dec 11 08:59:45 crc kubenswrapper[4629]: I1211 08:59:45.119407 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ad9572b-797d-4696-bd0a-2f6c388827ff","Type":"ContainerStarted","Data":"4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1"} Dec 11 08:59:46 crc kubenswrapper[4629]: I1211 08:59:46.136055 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ad9572b-797d-4696-bd0a-2f6c388827ff","Type":"ContainerStarted","Data":"7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9"} Dec 11 08:59:46 crc kubenswrapper[4629]: I1211 08:59:46.139711 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 08:59:46 crc kubenswrapper[4629]: I1211 08:59:46.186081 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.531978623 podStartE2EDuration="5.186062097s" podCreationTimestamp="2025-12-11 08:59:41 +0000 UTC" firstStartedPulling="2025-12-11 08:59:42.260429697 +0000 UTC m=+1370.325845305" lastFinishedPulling="2025-12-11 08:59:45.914513171 +0000 UTC m=+1373.979928779" observedRunningTime="2025-12-11 08:59:46.168779384 +0000 UTC m=+1374.234194992" watchObservedRunningTime="2025-12-11 08:59:46.186062097 +0000 UTC m=+1374.251477705" Dec 11 08:59:50 crc kubenswrapper[4629]: I1211 08:59:50.170666 4629 generic.go:334] "Generic (PLEG): container finished" podID="45f8baeb-687e-4c40-8ea9-05e66a26c9a7" containerID="91ffd72b0961ed31745502586d4d8ddfc25313b7cf1442f3495c7b21f64208c5" exitCode=0 Dec 11 08:59:50 crc kubenswrapper[4629]: I1211 08:59:50.170824 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-g6gr5" event={"ID":"45f8baeb-687e-4c40-8ea9-05e66a26c9a7","Type":"ContainerDied","Data":"91ffd72b0961ed31745502586d4d8ddfc25313b7cf1442f3495c7b21f64208c5"} Dec 11 08:59:50 crc kubenswrapper[4629]: I1211 08:59:50.486227 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:59:50 crc kubenswrapper[4629]: I1211 08:59:50.486314 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.505802 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.657150 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-scripts\") pod \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.657257 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-combined-ca-bundle\") pod \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.657319 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-config-data\") pod \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.657336 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqmjr\" (UniqueName: \"kubernetes.io/projected/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-kube-api-access-zqmjr\") pod \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\" (UID: \"45f8baeb-687e-4c40-8ea9-05e66a26c9a7\") " Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.663318 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-kube-api-access-zqmjr" (OuterVolumeSpecName: "kube-api-access-zqmjr") pod "45f8baeb-687e-4c40-8ea9-05e66a26c9a7" (UID: "45f8baeb-687e-4c40-8ea9-05e66a26c9a7"). InnerVolumeSpecName "kube-api-access-zqmjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.664225 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-scripts" (OuterVolumeSpecName: "scripts") pod "45f8baeb-687e-4c40-8ea9-05e66a26c9a7" (UID: "45f8baeb-687e-4c40-8ea9-05e66a26c9a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.690006 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-config-data" (OuterVolumeSpecName: "config-data") pod "45f8baeb-687e-4c40-8ea9-05e66a26c9a7" (UID: "45f8baeb-687e-4c40-8ea9-05e66a26c9a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.694333 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45f8baeb-687e-4c40-8ea9-05e66a26c9a7" (UID: "45f8baeb-687e-4c40-8ea9-05e66a26c9a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.759817 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.759868 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.759909 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:51 crc kubenswrapper[4629]: I1211 08:59:51.759917 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqmjr\" (UniqueName: \"kubernetes.io/projected/45f8baeb-687e-4c40-8ea9-05e66a26c9a7-kube-api-access-zqmjr\") on node \"crc\" DevicePath \"\"" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.191206 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-g6gr5" event={"ID":"45f8baeb-687e-4c40-8ea9-05e66a26c9a7","Type":"ContainerDied","Data":"b67187f47d38e912608f0a5e1069575b7341a197607cf00490a3cda2e3b33cd0"} Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.191253 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b67187f47d38e912608f0a5e1069575b7341a197607cf00490a3cda2e3b33cd0" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.191282 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-g6gr5" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.430723 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 11 08:59:52 crc kubenswrapper[4629]: E1211 08:59:52.431478 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45f8baeb-687e-4c40-8ea9-05e66a26c9a7" containerName="nova-cell0-conductor-db-sync" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.431502 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="45f8baeb-687e-4c40-8ea9-05e66a26c9a7" containerName="nova-cell0-conductor-db-sync" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.431689 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="45f8baeb-687e-4c40-8ea9-05e66a26c9a7" containerName="nova-cell0-conductor-db-sync" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.432433 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.437905 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.438134 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gb65z" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.447541 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.546114 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c34c47c-61ee-45e6-a289-5562eb5451e5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3c34c47c-61ee-45e6-a289-5562eb5451e5\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.546225 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c34c47c-61ee-45e6-a289-5562eb5451e5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3c34c47c-61ee-45e6-a289-5562eb5451e5\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.546374 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k657w\" (UniqueName: \"kubernetes.io/projected/3c34c47c-61ee-45e6-a289-5562eb5451e5-kube-api-access-k657w\") pod \"nova-cell0-conductor-0\" (UID: \"3c34c47c-61ee-45e6-a289-5562eb5451e5\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.647706 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c34c47c-61ee-45e6-a289-5562eb5451e5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3c34c47c-61ee-45e6-a289-5562eb5451e5\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.647799 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c34c47c-61ee-45e6-a289-5562eb5451e5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3c34c47c-61ee-45e6-a289-5562eb5451e5\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.647877 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k657w\" (UniqueName: \"kubernetes.io/projected/3c34c47c-61ee-45e6-a289-5562eb5451e5-kube-api-access-k657w\") pod \"nova-cell0-conductor-0\" (UID: \"3c34c47c-61ee-45e6-a289-5562eb5451e5\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.649992 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.659694 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c34c47c-61ee-45e6-a289-5562eb5451e5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3c34c47c-61ee-45e6-a289-5562eb5451e5\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.667714 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c34c47c-61ee-45e6-a289-5562eb5451e5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3c34c47c-61ee-45e6-a289-5562eb5451e5\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.670454 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k657w\" (UniqueName: \"kubernetes.io/projected/3c34c47c-61ee-45e6-a289-5562eb5451e5-kube-api-access-k657w\") pod \"nova-cell0-conductor-0\" (UID: \"3c34c47c-61ee-45e6-a289-5562eb5451e5\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.760776 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gb65z" Dec 11 08:59:52 crc kubenswrapper[4629]: I1211 08:59:52.769105 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 11 08:59:53 crc kubenswrapper[4629]: I1211 08:59:53.237804 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 11 08:59:54 crc kubenswrapper[4629]: I1211 08:59:54.240055 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 11 08:59:54 crc kubenswrapper[4629]: I1211 08:59:54.240407 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3c34c47c-61ee-45e6-a289-5562eb5451e5","Type":"ContainerStarted","Data":"8dd2b335f8a4c7e4f6e7a617dd93f7fae465b3fc8395c27952c6ea5f269b5216"} Dec 11 08:59:54 crc kubenswrapper[4629]: I1211 08:59:54.240433 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3c34c47c-61ee-45e6-a289-5562eb5451e5","Type":"ContainerStarted","Data":"6cb8e1c0045907689afb192a0cda6b77fc9dc30768059d553b9b07650460f9e3"} Dec 11 08:59:54 crc kubenswrapper[4629]: I1211 08:59:54.258723 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.258705344 podStartE2EDuration="2.258705344s" podCreationTimestamp="2025-12-11 08:59:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:59:54.252552521 +0000 UTC m=+1382.317968149" watchObservedRunningTime="2025-12-11 08:59:54.258705344 +0000 UTC m=+1382.324120952" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.151292 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np"] Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.156472 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.161205 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.161225 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.186771 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np"] Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.193101 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-config-volume\") pod \"collect-profiles-29424060-p65np\" (UID: \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.193166 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-secret-volume\") pod \"collect-profiles-29424060-p65np\" (UID: \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.193303 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv62g\" (UniqueName: \"kubernetes.io/projected/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-kube-api-access-sv62g\") pod \"collect-profiles-29424060-p65np\" (UID: \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.294803 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-config-volume\") pod \"collect-profiles-29424060-p65np\" (UID: \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.294898 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-secret-volume\") pod \"collect-profiles-29424060-p65np\" (UID: \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.294989 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv62g\" (UniqueName: \"kubernetes.io/projected/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-kube-api-access-sv62g\") pod \"collect-profiles-29424060-p65np\" (UID: \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.296097 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-config-volume\") pod \"collect-profiles-29424060-p65np\" (UID: \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.311657 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-secret-volume\") pod \"collect-profiles-29424060-p65np\" (UID: \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.315231 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv62g\" (UniqueName: \"kubernetes.io/projected/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-kube-api-access-sv62g\") pod \"collect-profiles-29424060-p65np\" (UID: \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.489226 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:00 crc kubenswrapper[4629]: I1211 09:00:00.979392 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np"] Dec 11 09:00:01 crc kubenswrapper[4629]: I1211 09:00:01.293046 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" event={"ID":"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7","Type":"ContainerStarted","Data":"ede2ebe08e6e45aa534e335a588bcd3cc44b55eb49b3b59956aaf1ccfb6207d0"} Dec 11 09:00:01 crc kubenswrapper[4629]: I1211 09:00:01.293770 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" event={"ID":"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7","Type":"ContainerStarted","Data":"b42e65a493b614f750ace55977fbbc0d7b36228e4b25e00af6338677a41e0c83"} Dec 11 09:00:01 crc kubenswrapper[4629]: I1211 09:00:01.318732 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" podStartSLOduration=1.31870928 podStartE2EDuration="1.31870928s" podCreationTimestamp="2025-12-11 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:01.318382149 +0000 UTC m=+1389.383797757" watchObservedRunningTime="2025-12-11 09:00:01.31870928 +0000 UTC m=+1389.384124888" Dec 11 09:00:02 crc kubenswrapper[4629]: I1211 09:00:02.305730 4629 generic.go:334] "Generic (PLEG): container finished" podID="94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7" containerID="ede2ebe08e6e45aa534e335a588bcd3cc44b55eb49b3b59956aaf1ccfb6207d0" exitCode=0 Dec 11 09:00:02 crc kubenswrapper[4629]: I1211 09:00:02.305780 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" event={"ID":"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7","Type":"ContainerDied","Data":"ede2ebe08e6e45aa534e335a588bcd3cc44b55eb49b3b59956aaf1ccfb6207d0"} Dec 11 09:00:02 crc kubenswrapper[4629]: I1211 09:00:02.794503 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.334732 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-vppnj"] Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.336302 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.342495 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.346330 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.353229 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vppnj"] Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.363475 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vppnj\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.363648 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4hrq\" (UniqueName: \"kubernetes.io/projected/a5f5f734-35d7-49f9-a007-d1722d4d4673-kube-api-access-g4hrq\") pod \"nova-cell0-cell-mapping-vppnj\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.363726 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-scripts\") pod \"nova-cell0-cell-mapping-vppnj\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.363807 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-config-data\") pod \"nova-cell0-cell-mapping-vppnj\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.465812 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vppnj\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.466069 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4hrq\" (UniqueName: \"kubernetes.io/projected/a5f5f734-35d7-49f9-a007-d1722d4d4673-kube-api-access-g4hrq\") pod \"nova-cell0-cell-mapping-vppnj\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.466116 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-scripts\") pod \"nova-cell0-cell-mapping-vppnj\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.466164 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-config-data\") pod \"nova-cell0-cell-mapping-vppnj\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.496321 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-config-data\") pod \"nova-cell0-cell-mapping-vppnj\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.497383 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vppnj\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.501305 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-scripts\") pod \"nova-cell0-cell-mapping-vppnj\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.516650 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4hrq\" (UniqueName: \"kubernetes.io/projected/a5f5f734-35d7-49f9-a007-d1722d4d4673-kube-api-access-g4hrq\") pod \"nova-cell0-cell-mapping-vppnj\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.544544 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.547214 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.557033 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.566637 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpv55\" (UniqueName: \"kubernetes.io/projected/64a3e1bc-8fa8-4493-ac7c-865e51169162-kube-api-access-kpv55\") pod \"nova-api-0\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.566679 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64a3e1bc-8fa8-4493-ac7c-865e51169162-config-data\") pod \"nova-api-0\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.566711 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64a3e1bc-8fa8-4493-ac7c-865e51169162-logs\") pod \"nova-api-0\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.566746 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64a3e1bc-8fa8-4493-ac7c-865e51169162-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.577357 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.683833 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.685408 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpv55\" (UniqueName: \"kubernetes.io/projected/64a3e1bc-8fa8-4493-ac7c-865e51169162-kube-api-access-kpv55\") pod \"nova-api-0\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.685450 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64a3e1bc-8fa8-4493-ac7c-865e51169162-config-data\") pod \"nova-api-0\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.685486 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64a3e1bc-8fa8-4493-ac7c-865e51169162-logs\") pod \"nova-api-0\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.685540 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64a3e1bc-8fa8-4493-ac7c-865e51169162-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.703511 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64a3e1bc-8fa8-4493-ac7c-865e51169162-logs\") pod \"nova-api-0\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.731795 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.734105 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.737719 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64a3e1bc-8fa8-4493-ac7c-865e51169162-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.752566 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.775820 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64a3e1bc-8fa8-4493-ac7c-865e51169162-config-data\") pod \"nova-api-0\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.793635 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpv55\" (UniqueName: \"kubernetes.io/projected/64a3e1bc-8fa8-4493-ac7c-865e51169162-kube-api-access-kpv55\") pod \"nova-api-0\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.820048 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf970abd-834e-4a87-9503-f049ccf7d4fe-logs\") pod \"nova-metadata-0\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " pod="openstack/nova-metadata-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.820435 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf970abd-834e-4a87-9503-f049ccf7d4fe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " pod="openstack/nova-metadata-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.820658 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf970abd-834e-4a87-9503-f049ccf7d4fe-config-data\") pod \"nova-metadata-0\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " pod="openstack/nova-metadata-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.820753 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klmz7\" (UniqueName: \"kubernetes.io/projected/bf970abd-834e-4a87-9503-f049ccf7d4fe-kube-api-access-klmz7\") pod \"nova-metadata-0\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " pod="openstack/nova-metadata-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.829553 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.847925 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.849206 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.873704 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.891133 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.960251 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf970abd-834e-4a87-9503-f049ccf7d4fe-logs\") pod \"nova-metadata-0\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " pod="openstack/nova-metadata-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.963419 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf970abd-834e-4a87-9503-f049ccf7d4fe-logs\") pod \"nova-metadata-0\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " pod="openstack/nova-metadata-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.969675 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8336f4a-02d9-4423-b40f-231ee4781b89-config-data\") pod \"nova-scheduler-0\" (UID: \"a8336f4a-02d9-4423-b40f-231ee4781b89\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.969774 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l7x6\" (UniqueName: \"kubernetes.io/projected/a8336f4a-02d9-4423-b40f-231ee4781b89-kube-api-access-2l7x6\") pod \"nova-scheduler-0\" (UID: \"a8336f4a-02d9-4423-b40f-231ee4781b89\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.969876 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf970abd-834e-4a87-9503-f049ccf7d4fe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " pod="openstack/nova-metadata-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.970055 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf970abd-834e-4a87-9503-f049ccf7d4fe-config-data\") pod \"nova-metadata-0\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " pod="openstack/nova-metadata-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.970109 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klmz7\" (UniqueName: \"kubernetes.io/projected/bf970abd-834e-4a87-9503-f049ccf7d4fe-kube-api-access-klmz7\") pod \"nova-metadata-0\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " pod="openstack/nova-metadata-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.970169 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8336f4a-02d9-4423-b40f-231ee4781b89-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a8336f4a-02d9-4423-b40f-231ee4781b89\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.974164 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf970abd-834e-4a87-9503-f049ccf7d4fe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " pod="openstack/nova-metadata-0" Dec 11 09:00:03 crc kubenswrapper[4629]: I1211 09:00:03.984252 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.020297 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klmz7\" (UniqueName: \"kubernetes.io/projected/bf970abd-834e-4a87-9503-f049ccf7d4fe-kube-api-access-klmz7\") pod \"nova-metadata-0\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " pod="openstack/nova-metadata-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.026321 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf970abd-834e-4a87-9503-f049ccf7d4fe-config-data\") pod \"nova-metadata-0\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " pod="openstack/nova-metadata-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.074566 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-55m6h"] Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.076567 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.079298 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8336f4a-02d9-4423-b40f-231ee4781b89-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a8336f4a-02d9-4423-b40f-231ee4781b89\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.079389 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8336f4a-02d9-4423-b40f-231ee4781b89-config-data\") pod \"nova-scheduler-0\" (UID: \"a8336f4a-02d9-4423-b40f-231ee4781b89\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.079413 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l7x6\" (UniqueName: \"kubernetes.io/projected/a8336f4a-02d9-4423-b40f-231ee4781b89-kube-api-access-2l7x6\") pod \"nova-scheduler-0\" (UID: \"a8336f4a-02d9-4423-b40f-231ee4781b89\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.092609 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8336f4a-02d9-4423-b40f-231ee4781b89-config-data\") pod \"nova-scheduler-0\" (UID: \"a8336f4a-02d9-4423-b40f-231ee4781b89\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.097750 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8336f4a-02d9-4423-b40f-231ee4781b89-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a8336f4a-02d9-4423-b40f-231ee4781b89\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.124252 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l7x6\" (UniqueName: \"kubernetes.io/projected/a8336f4a-02d9-4423-b40f-231ee4781b89-kube-api-access-2l7x6\") pod \"nova-scheduler-0\" (UID: \"a8336f4a-02d9-4423-b40f-231ee4781b89\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.125266 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-55m6h"] Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.160931 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.161260 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.161621 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7" containerName="collect-profiles" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.162386 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.164435 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.180712 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.180777 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q5n2\" (UniqueName: \"kubernetes.io/projected/5c29fe70-580c-4f60-9dbc-d9bae34e376d-kube-api-access-7q5n2\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.180856 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-config\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.180894 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.180932 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.183583 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.197823 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.220400 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.283078 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-secret-volume\") pod \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\" (UID: \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\") " Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.283141 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv62g\" (UniqueName: \"kubernetes.io/projected/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-kube-api-access-sv62g\") pod \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\" (UID: \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\") " Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.283173 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-config-volume\") pod \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\" (UID: \"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7\") " Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.283471 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-config\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.283514 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.283541 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.283578 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl29k\" (UniqueName: \"kubernetes.io/projected/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-kube-api-access-bl29k\") pod \"nova-cell1-novncproxy-0\" (UID: \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.283615 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.283735 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.283777 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q5n2\" (UniqueName: \"kubernetes.io/projected/5c29fe70-580c-4f60-9dbc-d9bae34e376d-kube-api-access-7q5n2\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.283809 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.285260 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-config\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.288177 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.291636 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-config-volume" (OuterVolumeSpecName: "config-volume") pod "94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7" (UID: "94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.293454 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.294423 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.297099 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-kube-api-access-sv62g" (OuterVolumeSpecName: "kube-api-access-sv62g") pod "94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7" (UID: "94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7"). InnerVolumeSpecName "kube-api-access-sv62g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.297105 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7" (UID: "94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.318405 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q5n2\" (UniqueName: \"kubernetes.io/projected/5c29fe70-580c-4f60-9dbc-d9bae34e376d-kube-api-access-7q5n2\") pod \"dnsmasq-dns-8b8cf6657-55m6h\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.335474 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" event={"ID":"94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7","Type":"ContainerDied","Data":"b42e65a493b614f750ace55977fbbc0d7b36228e4b25e00af6338677a41e0c83"} Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.335528 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b42e65a493b614f750ace55977fbbc0d7b36228e4b25e00af6338677a41e0c83" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.335546 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-p65np" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.389000 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.389107 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.389160 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl29k\" (UniqueName: \"kubernetes.io/projected/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-kube-api-access-bl29k\") pod \"nova-cell1-novncproxy-0\" (UID: \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.389303 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv62g\" (UniqueName: \"kubernetes.io/projected/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-kube-api-access-sv62g\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.389325 4629 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.389339 4629 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.413972 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.416450 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.421298 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.427129 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl29k\" (UniqueName: \"kubernetes.io/projected/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-kube-api-access-bl29k\") pod \"nova-cell1-novncproxy-0\" (UID: \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.504241 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:04 crc kubenswrapper[4629]: E1211 09:00:04.665056 4629 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94db49ab_8283_4f02_b9e9_8d0bc5e0d5f7.slice/crio-b42e65a493b614f750ace55977fbbc0d7b36228e4b25e00af6338677a41e0c83\": RecentStats: unable to find data in memory cache]" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.667985 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vppnj"] Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.895060 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.994053 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-65bhx"] Dec 11 09:00:04 crc kubenswrapper[4629]: E1211 09:00:04.994764 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7" containerName="collect-profiles" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.994782 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="94db49ab-8283-4f02-b9e9-8d0bc5e0d5f7" containerName="collect-profiles" Dec 11 09:00:04 crc kubenswrapper[4629]: I1211 09:00:04.995699 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.011758 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.011970 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.033750 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-65bhx"] Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.114677 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-65bhx\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.114860 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-config-data\") pod \"nova-cell1-conductor-db-sync-65bhx\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.115065 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-scripts\") pod \"nova-cell1-conductor-db-sync-65bhx\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.115096 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhw6z\" (UniqueName: \"kubernetes.io/projected/8f83bf9d-806c-4310-9315-2bcbd529a6d1-kube-api-access-xhw6z\") pod \"nova-cell1-conductor-db-sync-65bhx\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.154256 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.222002 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-scripts\") pod \"nova-cell1-conductor-db-sync-65bhx\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.222056 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhw6z\" (UniqueName: \"kubernetes.io/projected/8f83bf9d-806c-4310-9315-2bcbd529a6d1-kube-api-access-xhw6z\") pod \"nova-cell1-conductor-db-sync-65bhx\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.222128 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-65bhx\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.222216 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-config-data\") pod \"nova-cell1-conductor-db-sync-65bhx\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.228772 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-config-data\") pod \"nova-cell1-conductor-db-sync-65bhx\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.229303 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-65bhx\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.229373 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-scripts\") pod \"nova-cell1-conductor-db-sync-65bhx\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.247113 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhw6z\" (UniqueName: \"kubernetes.io/projected/8f83bf9d-806c-4310-9315-2bcbd529a6d1-kube-api-access-xhw6z\") pod \"nova-cell1-conductor-db-sync-65bhx\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.348300 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.359126 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a8336f4a-02d9-4423-b40f-231ee4781b89","Type":"ContainerStarted","Data":"0d9119215206da99d1199e4762f08f333997526d94d0d017a78fd4655afe5b20"} Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.368207 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vppnj" event={"ID":"a5f5f734-35d7-49f9-a007-d1722d4d4673","Type":"ContainerStarted","Data":"a42cfdb3a956accc8e6e35bf29d2aa5681ece9adba1c92efc77474b2a3fb1f98"} Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.368262 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vppnj" event={"ID":"a5f5f734-35d7-49f9-a007-d1722d4d4673","Type":"ContainerStarted","Data":"ef33646d470b46b86370961feef7dede3ee7fbdff54fa7b41260db0729f3c949"} Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.375111 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64a3e1bc-8fa8-4493-ac7c-865e51169162","Type":"ContainerStarted","Data":"5f09c59420a9db9f9a4d10fcd99c4913288c4bf0a0920fb4c7ba739336791714"} Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.378396 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-55m6h"] Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.635110 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-vppnj" podStartSLOduration=2.635084106 podStartE2EDuration="2.635084106s" podCreationTimestamp="2025-12-11 09:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:05.411735644 +0000 UTC m=+1393.477151262" watchObservedRunningTime="2025-12-11 09:00:05.635084106 +0000 UTC m=+1393.700499714" Dec 11 09:00:05 crc kubenswrapper[4629]: W1211 09:00:05.639245 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf970abd_834e_4a87_9503_f049ccf7d4fe.slice/crio-c71fe01437773a41a3200fdcca1b0b74b088aac8f1e02845161f5fc7dd7ebe4a WatchSource:0}: Error finding container c71fe01437773a41a3200fdcca1b0b74b088aac8f1e02845161f5fc7dd7ebe4a: Status 404 returned error can't find the container with id c71fe01437773a41a3200fdcca1b0b74b088aac8f1e02845161f5fc7dd7ebe4a Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.641617 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.701365 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:05 crc kubenswrapper[4629]: W1211 09:00:05.726342 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3204c2e_39a2_4fa3_a42e_33aa54521f0e.slice/crio-68ba5b13116bd9e5ac20002414b0ea46735c3eace4f9c99621a285dbccd06abf WatchSource:0}: Error finding container 68ba5b13116bd9e5ac20002414b0ea46735c3eace4f9c99621a285dbccd06abf: Status 404 returned error can't find the container with id 68ba5b13116bd9e5ac20002414b0ea46735c3eace4f9c99621a285dbccd06abf Dec 11 09:00:05 crc kubenswrapper[4629]: I1211 09:00:05.957752 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-65bhx"] Dec 11 09:00:06 crc kubenswrapper[4629]: I1211 09:00:06.385892 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b3204c2e-39a2-4fa3-a42e-33aa54521f0e","Type":"ContainerStarted","Data":"68ba5b13116bd9e5ac20002414b0ea46735c3eace4f9c99621a285dbccd06abf"} Dec 11 09:00:06 crc kubenswrapper[4629]: I1211 09:00:06.388509 4629 generic.go:334] "Generic (PLEG): container finished" podID="5c29fe70-580c-4f60-9dbc-d9bae34e376d" containerID="d3059bb274a5022af71e9262aa740e8c6cb9ad05b76b83e91a430312f6cca7e7" exitCode=0 Dec 11 09:00:06 crc kubenswrapper[4629]: I1211 09:00:06.388605 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" event={"ID":"5c29fe70-580c-4f60-9dbc-d9bae34e376d","Type":"ContainerDied","Data":"d3059bb274a5022af71e9262aa740e8c6cb9ad05b76b83e91a430312f6cca7e7"} Dec 11 09:00:06 crc kubenswrapper[4629]: I1211 09:00:06.388654 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" event={"ID":"5c29fe70-580c-4f60-9dbc-d9bae34e376d","Type":"ContainerStarted","Data":"4accc909aee8920af1e2c3ece6d4d8ae07646abbdf0bbbe7fb8fc16118ca9bc6"} Dec 11 09:00:06 crc kubenswrapper[4629]: I1211 09:00:06.394271 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-65bhx" event={"ID":"8f83bf9d-806c-4310-9315-2bcbd529a6d1","Type":"ContainerStarted","Data":"9f9ffba2e0fd7c28249813b71c9e26135de39e68728e184c968d7475561fa5ba"} Dec 11 09:00:06 crc kubenswrapper[4629]: I1211 09:00:06.394395 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-65bhx" event={"ID":"8f83bf9d-806c-4310-9315-2bcbd529a6d1","Type":"ContainerStarted","Data":"6c90fcb38758fe12e6b0f10c34bdf81935e886ea357bb959da463f5e1ad9401b"} Dec 11 09:00:06 crc kubenswrapper[4629]: I1211 09:00:06.396774 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf970abd-834e-4a87-9503-f049ccf7d4fe","Type":"ContainerStarted","Data":"c71fe01437773a41a3200fdcca1b0b74b088aac8f1e02845161f5fc7dd7ebe4a"} Dec 11 09:00:06 crc kubenswrapper[4629]: I1211 09:00:06.442385 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-65bhx" podStartSLOduration=2.442363111 podStartE2EDuration="2.442363111s" podCreationTimestamp="2025-12-11 09:00:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:06.437092256 +0000 UTC m=+1394.502507884" watchObservedRunningTime="2025-12-11 09:00:06.442363111 +0000 UTC m=+1394.507778719" Dec 11 09:00:07 crc kubenswrapper[4629]: I1211 09:00:07.439200 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" event={"ID":"5c29fe70-580c-4f60-9dbc-d9bae34e376d","Type":"ContainerStarted","Data":"8cdd2e499d8a8d46810c4f6a3b338bc25758cb8545afb5aca32df8a634eabb75"} Dec 11 09:00:07 crc kubenswrapper[4629]: I1211 09:00:07.470011 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" podStartSLOduration=4.469988064 podStartE2EDuration="4.469988064s" podCreationTimestamp="2025-12-11 09:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:07.468823047 +0000 UTC m=+1395.534238665" watchObservedRunningTime="2025-12-11 09:00:07.469988064 +0000 UTC m=+1395.535403682" Dec 11 09:00:08 crc kubenswrapper[4629]: I1211 09:00:08.100583 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:08 crc kubenswrapper[4629]: I1211 09:00:08.135374 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 09:00:08 crc kubenswrapper[4629]: I1211 09:00:08.453999 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:10 crc kubenswrapper[4629]: I1211 09:00:10.474576 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a8336f4a-02d9-4423-b40f-231ee4781b89","Type":"ContainerStarted","Data":"345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9"} Dec 11 09:00:10 crc kubenswrapper[4629]: I1211 09:00:10.477728 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64a3e1bc-8fa8-4493-ac7c-865e51169162","Type":"ContainerStarted","Data":"afa498f6abcebe5d64954bd9a552c17c9ff1a92839177a5c95be94fecc0e3ef9"} Dec 11 09:00:10 crc kubenswrapper[4629]: I1211 09:00:10.480283 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b3204c2e-39a2-4fa3-a42e-33aa54521f0e","Type":"ContainerStarted","Data":"d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3"} Dec 11 09:00:10 crc kubenswrapper[4629]: I1211 09:00:10.480476 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="b3204c2e-39a2-4fa3-a42e-33aa54521f0e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3" gracePeriod=30 Dec 11 09:00:10 crc kubenswrapper[4629]: I1211 09:00:10.486506 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf970abd-834e-4a87-9503-f049ccf7d4fe","Type":"ContainerStarted","Data":"55ccfd39f649028747aa055dac46812f5680c426715ef70823ae69026dd1be45"} Dec 11 09:00:10 crc kubenswrapper[4629]: I1211 09:00:10.501923 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.705679684 podStartE2EDuration="7.501904854s" podCreationTimestamp="2025-12-11 09:00:03 +0000 UTC" firstStartedPulling="2025-12-11 09:00:05.176970233 +0000 UTC m=+1393.242385831" lastFinishedPulling="2025-12-11 09:00:09.973195393 +0000 UTC m=+1398.038611001" observedRunningTime="2025-12-11 09:00:10.499552979 +0000 UTC m=+1398.564968597" watchObservedRunningTime="2025-12-11 09:00:10.501904854 +0000 UTC m=+1398.567320462" Dec 11 09:00:10 crc kubenswrapper[4629]: I1211 09:00:10.537119 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.292997991 podStartE2EDuration="7.537096328s" podCreationTimestamp="2025-12-11 09:00:03 +0000 UTC" firstStartedPulling="2025-12-11 09:00:05.74129443 +0000 UTC m=+1393.806710038" lastFinishedPulling="2025-12-11 09:00:09.985392757 +0000 UTC m=+1398.050808375" observedRunningTime="2025-12-11 09:00:10.522570362 +0000 UTC m=+1398.587985990" watchObservedRunningTime="2025-12-11 09:00:10.537096328 +0000 UTC m=+1398.602511936" Dec 11 09:00:11 crc kubenswrapper[4629]: I1211 09:00:11.500466 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf970abd-834e-4a87-9503-f049ccf7d4fe","Type":"ContainerStarted","Data":"ab219aff897df71f45886e108ac4bc8809734b3f7530a489ef8bcdfd4855b34e"} Dec 11 09:00:11 crc kubenswrapper[4629]: I1211 09:00:11.500615 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bf970abd-834e-4a87-9503-f049ccf7d4fe" containerName="nova-metadata-log" containerID="cri-o://55ccfd39f649028747aa055dac46812f5680c426715ef70823ae69026dd1be45" gracePeriod=30 Dec 11 09:00:11 crc kubenswrapper[4629]: I1211 09:00:11.500785 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bf970abd-834e-4a87-9503-f049ccf7d4fe" containerName="nova-metadata-metadata" containerID="cri-o://ab219aff897df71f45886e108ac4bc8809734b3f7530a489ef8bcdfd4855b34e" gracePeriod=30 Dec 11 09:00:11 crc kubenswrapper[4629]: I1211 09:00:11.504870 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64a3e1bc-8fa8-4493-ac7c-865e51169162","Type":"ContainerStarted","Data":"ba34b4bae6ba99fb9c0b068508d3ec43556b0216a99b57085667ebc210a1e399"} Dec 11 09:00:11 crc kubenswrapper[4629]: I1211 09:00:11.530753 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.198326904 podStartE2EDuration="8.530730094s" podCreationTimestamp="2025-12-11 09:00:03 +0000 UTC" firstStartedPulling="2025-12-11 09:00:05.642124466 +0000 UTC m=+1393.707540074" lastFinishedPulling="2025-12-11 09:00:09.974527666 +0000 UTC m=+1398.039943264" observedRunningTime="2025-12-11 09:00:11.527697689 +0000 UTC m=+1399.593113327" watchObservedRunningTime="2025-12-11 09:00:11.530730094 +0000 UTC m=+1399.596145712" Dec 11 09:00:11 crc kubenswrapper[4629]: I1211 09:00:11.579384 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.547844604 podStartE2EDuration="8.579363271s" podCreationTimestamp="2025-12-11 09:00:03 +0000 UTC" firstStartedPulling="2025-12-11 09:00:04.933237251 +0000 UTC m=+1392.998652859" lastFinishedPulling="2025-12-11 09:00:09.964755918 +0000 UTC m=+1398.030171526" observedRunningTime="2025-12-11 09:00:11.573780356 +0000 UTC m=+1399.639195964" watchObservedRunningTime="2025-12-11 09:00:11.579363271 +0000 UTC m=+1399.644778879" Dec 11 09:00:11 crc kubenswrapper[4629]: I1211 09:00:11.772138 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 11 09:00:12 crc kubenswrapper[4629]: I1211 09:00:12.515652 4629 generic.go:334] "Generic (PLEG): container finished" podID="bf970abd-834e-4a87-9503-f049ccf7d4fe" containerID="ab219aff897df71f45886e108ac4bc8809734b3f7530a489ef8bcdfd4855b34e" exitCode=0 Dec 11 09:00:12 crc kubenswrapper[4629]: I1211 09:00:12.516033 4629 generic.go:334] "Generic (PLEG): container finished" podID="bf970abd-834e-4a87-9503-f049ccf7d4fe" containerID="55ccfd39f649028747aa055dac46812f5680c426715ef70823ae69026dd1be45" exitCode=143 Dec 11 09:00:12 crc kubenswrapper[4629]: I1211 09:00:12.515725 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf970abd-834e-4a87-9503-f049ccf7d4fe","Type":"ContainerDied","Data":"ab219aff897df71f45886e108ac4bc8809734b3f7530a489ef8bcdfd4855b34e"} Dec 11 09:00:12 crc kubenswrapper[4629]: I1211 09:00:12.516130 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf970abd-834e-4a87-9503-f049ccf7d4fe","Type":"ContainerDied","Data":"55ccfd39f649028747aa055dac46812f5680c426715ef70823ae69026dd1be45"} Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.033222 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.101651 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf970abd-834e-4a87-9503-f049ccf7d4fe-config-data\") pod \"bf970abd-834e-4a87-9503-f049ccf7d4fe\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.104997 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf970abd-834e-4a87-9503-f049ccf7d4fe-combined-ca-bundle\") pod \"bf970abd-834e-4a87-9503-f049ccf7d4fe\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.105254 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klmz7\" (UniqueName: \"kubernetes.io/projected/bf970abd-834e-4a87-9503-f049ccf7d4fe-kube-api-access-klmz7\") pod \"bf970abd-834e-4a87-9503-f049ccf7d4fe\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.105467 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf970abd-834e-4a87-9503-f049ccf7d4fe-logs\") pod \"bf970abd-834e-4a87-9503-f049ccf7d4fe\" (UID: \"bf970abd-834e-4a87-9503-f049ccf7d4fe\") " Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.110305 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf970abd-834e-4a87-9503-f049ccf7d4fe-logs" (OuterVolumeSpecName: "logs") pod "bf970abd-834e-4a87-9503-f049ccf7d4fe" (UID: "bf970abd-834e-4a87-9503-f049ccf7d4fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.135133 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf970abd-834e-4a87-9503-f049ccf7d4fe-kube-api-access-klmz7" (OuterVolumeSpecName: "kube-api-access-klmz7") pod "bf970abd-834e-4a87-9503-f049ccf7d4fe" (UID: "bf970abd-834e-4a87-9503-f049ccf7d4fe"). InnerVolumeSpecName "kube-api-access-klmz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.156311 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf970abd-834e-4a87-9503-f049ccf7d4fe-config-data" (OuterVolumeSpecName: "config-data") pod "bf970abd-834e-4a87-9503-f049ccf7d4fe" (UID: "bf970abd-834e-4a87-9503-f049ccf7d4fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.170311 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf970abd-834e-4a87-9503-f049ccf7d4fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf970abd-834e-4a87-9503-f049ccf7d4fe" (UID: "bf970abd-834e-4a87-9503-f049ccf7d4fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.208234 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klmz7\" (UniqueName: \"kubernetes.io/projected/bf970abd-834e-4a87-9503-f049ccf7d4fe-kube-api-access-klmz7\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.208270 4629 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf970abd-834e-4a87-9503-f049ccf7d4fe-logs\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.208282 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf970abd-834e-4a87-9503-f049ccf7d4fe-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.208291 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf970abd-834e-4a87-9503-f049ccf7d4fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.527580 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf970abd-834e-4a87-9503-f049ccf7d4fe","Type":"ContainerDied","Data":"c71fe01437773a41a3200fdcca1b0b74b088aac8f1e02845161f5fc7dd7ebe4a"} Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.527641 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.528995 4629 scope.go:117] "RemoveContainer" containerID="ab219aff897df71f45886e108ac4bc8809734b3f7530a489ef8bcdfd4855b34e" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.567261 4629 scope.go:117] "RemoveContainer" containerID="55ccfd39f649028747aa055dac46812f5680c426715ef70823ae69026dd1be45" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.575681 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.582254 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.599822 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:13 crc kubenswrapper[4629]: E1211 09:00:13.600322 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf970abd-834e-4a87-9503-f049ccf7d4fe" containerName="nova-metadata-metadata" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.600339 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf970abd-834e-4a87-9503-f049ccf7d4fe" containerName="nova-metadata-metadata" Dec 11 09:00:13 crc kubenswrapper[4629]: E1211 09:00:13.600356 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf970abd-834e-4a87-9503-f049ccf7d4fe" containerName="nova-metadata-log" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.600363 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf970abd-834e-4a87-9503-f049ccf7d4fe" containerName="nova-metadata-log" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.600587 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf970abd-834e-4a87-9503-f049ccf7d4fe" containerName="nova-metadata-metadata" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.600612 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf970abd-834e-4a87-9503-f049ccf7d4fe" containerName="nova-metadata-log" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.603270 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.608642 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.608899 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.640394 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.760059 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.760196 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqbqp\" (UniqueName: \"kubernetes.io/projected/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-kube-api-access-kqbqp\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.760236 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-config-data\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.760266 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-logs\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.760297 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.862361 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.862422 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqbqp\" (UniqueName: \"kubernetes.io/projected/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-kube-api-access-kqbqp\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.862460 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-config-data\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.862491 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-logs\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.862519 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.863202 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-logs\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.866511 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.869270 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.870729 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-config-data\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.879931 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqbqp\" (UniqueName: \"kubernetes.io/projected/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-kube-api-access-kqbqp\") pod \"nova-metadata-0\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " pod="openstack/nova-metadata-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.891787 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.891879 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 09:00:13 crc kubenswrapper[4629]: I1211 09:00:13.928130 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.210533 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf970abd-834e-4a87-9503-f049ccf7d4fe" path="/var/lib/kubelet/pods/bf970abd-834e-4a87-9503-f049ccf7d4fe/volumes" Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.221290 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.223039 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.269325 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.424044 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.437174 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.504866 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.538910 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-jcpxn"] Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.540238 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" podUID="a5290303-73ff-47c7-8b4e-6518841d75ec" containerName="dnsmasq-dns" containerID="cri-o://f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09" gracePeriod=10 Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.577375 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e708d4fe-3394-42cf-a1b8-e93804c8fd6d","Type":"ContainerStarted","Data":"7eea11652b7ef26d4cdec084beea387d04cda184c63a6ca0bb26cffbe7820156"} Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.633769 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.966299 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="64a3e1bc-8fa8-4493-ac7c-865e51169162" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.166:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 09:00:14 crc kubenswrapper[4629]: I1211 09:00:14.967327 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="64a3e1bc-8fa8-4493-ac7c-865e51169162" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.166:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.228746 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.229383 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="2594a86c-caf2-4874-a415-4e6fda88db08" containerName="kube-state-metrics" containerID="cri-o://1cd05b725be1d7583bf973ed8dba0ef5f55f3f153bef4abf0bc35e8be720622c" gracePeriod=30 Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.248374 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.420612 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-687zr\" (UniqueName: \"kubernetes.io/projected/a5290303-73ff-47c7-8b4e-6518841d75ec-kube-api-access-687zr\") pod \"a5290303-73ff-47c7-8b4e-6518841d75ec\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.420955 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-ovsdbserver-sb\") pod \"a5290303-73ff-47c7-8b4e-6518841d75ec\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.421099 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-config\") pod \"a5290303-73ff-47c7-8b4e-6518841d75ec\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.421176 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-ovsdbserver-nb\") pod \"a5290303-73ff-47c7-8b4e-6518841d75ec\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.421218 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-dns-svc\") pod \"a5290303-73ff-47c7-8b4e-6518841d75ec\" (UID: \"a5290303-73ff-47c7-8b4e-6518841d75ec\") " Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.426140 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5290303-73ff-47c7-8b4e-6518841d75ec-kube-api-access-687zr" (OuterVolumeSpecName: "kube-api-access-687zr") pod "a5290303-73ff-47c7-8b4e-6518841d75ec" (UID: "a5290303-73ff-47c7-8b4e-6518841d75ec"). InnerVolumeSpecName "kube-api-access-687zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.519026 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-config" (OuterVolumeSpecName: "config") pod "a5290303-73ff-47c7-8b4e-6518841d75ec" (UID: "a5290303-73ff-47c7-8b4e-6518841d75ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.524410 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-config\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.524442 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-687zr\" (UniqueName: \"kubernetes.io/projected/a5290303-73ff-47c7-8b4e-6518841d75ec-kube-api-access-687zr\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.525454 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a5290303-73ff-47c7-8b4e-6518841d75ec" (UID: "a5290303-73ff-47c7-8b4e-6518841d75ec"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.527461 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a5290303-73ff-47c7-8b4e-6518841d75ec" (UID: "a5290303-73ff-47c7-8b4e-6518841d75ec"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.565672 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a5290303-73ff-47c7-8b4e-6518841d75ec" (UID: "a5290303-73ff-47c7-8b4e-6518841d75ec"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.598268 4629 generic.go:334] "Generic (PLEG): container finished" podID="a5290303-73ff-47c7-8b4e-6518841d75ec" containerID="f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09" exitCode=0 Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.598361 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" event={"ID":"a5290303-73ff-47c7-8b4e-6518841d75ec","Type":"ContainerDied","Data":"f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09"} Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.598395 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" event={"ID":"a5290303-73ff-47c7-8b4e-6518841d75ec","Type":"ContainerDied","Data":"c868fdbea4ccd1dfe4ed23ddc98bf42e455fa58aadeeb8e42cc56790aa55b7ef"} Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.598396 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-jcpxn" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.598414 4629 scope.go:117] "RemoveContainer" containerID="f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.608484 4629 generic.go:334] "Generic (PLEG): container finished" podID="2594a86c-caf2-4874-a415-4e6fda88db08" containerID="1cd05b725be1d7583bf973ed8dba0ef5f55f3f153bef4abf0bc35e8be720622c" exitCode=2 Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.608561 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2594a86c-caf2-4874-a415-4e6fda88db08","Type":"ContainerDied","Data":"1cd05b725be1d7583bf973ed8dba0ef5f55f3f153bef4abf0bc35e8be720622c"} Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.614044 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e708d4fe-3394-42cf-a1b8-e93804c8fd6d","Type":"ContainerStarted","Data":"cc07115752980f532522f711989a142b1be931d1c2dc347dd40a23d2e15d5880"} Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.614091 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e708d4fe-3394-42cf-a1b8-e93804c8fd6d","Type":"ContainerStarted","Data":"b1392b230a8cb39edecd1065c18563af1246b312d49cea2b5d533cc90bec0a0e"} Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.627659 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.627724 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.627741 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5290303-73ff-47c7-8b4e-6518841d75ec-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.660305 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-jcpxn"] Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.676080 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-jcpxn"] Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.708194 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.7081743400000002 podStartE2EDuration="2.70817434s" podCreationTimestamp="2025-12-11 09:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:15.677323372 +0000 UTC m=+1403.742738980" watchObservedRunningTime="2025-12-11 09:00:15.70817434 +0000 UTC m=+1403.773589948" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.712322 4629 scope.go:117] "RemoveContainer" containerID="968141be1aa24ffb4d8165374fb571fe86f434479c72211bb48d5b7fcab78fa3" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.772595 4629 scope.go:117] "RemoveContainer" containerID="f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09" Dec 11 09:00:15 crc kubenswrapper[4629]: E1211 09:00:15.775033 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09\": container with ID starting with f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09 not found: ID does not exist" containerID="f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.775094 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09"} err="failed to get container status \"f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09\": rpc error: code = NotFound desc = could not find container \"f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09\": container with ID starting with f6c2f22e85e34b492632377d00465393626a53a083710e20c85e7d87fdd95e09 not found: ID does not exist" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.775126 4629 scope.go:117] "RemoveContainer" containerID="968141be1aa24ffb4d8165374fb571fe86f434479c72211bb48d5b7fcab78fa3" Dec 11 09:00:15 crc kubenswrapper[4629]: E1211 09:00:15.780000 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"968141be1aa24ffb4d8165374fb571fe86f434479c72211bb48d5b7fcab78fa3\": container with ID starting with 968141be1aa24ffb4d8165374fb571fe86f434479c72211bb48d5b7fcab78fa3 not found: ID does not exist" containerID="968141be1aa24ffb4d8165374fb571fe86f434479c72211bb48d5b7fcab78fa3" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.780047 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"968141be1aa24ffb4d8165374fb571fe86f434479c72211bb48d5b7fcab78fa3"} err="failed to get container status \"968141be1aa24ffb4d8165374fb571fe86f434479c72211bb48d5b7fcab78fa3\": rpc error: code = NotFound desc = could not find container \"968141be1aa24ffb4d8165374fb571fe86f434479c72211bb48d5b7fcab78fa3\": container with ID starting with 968141be1aa24ffb4d8165374fb571fe86f434479c72211bb48d5b7fcab78fa3 not found: ID does not exist" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.806656 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.935710 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8jnv\" (UniqueName: \"kubernetes.io/projected/2594a86c-caf2-4874-a415-4e6fda88db08-kube-api-access-d8jnv\") pod \"2594a86c-caf2-4874-a415-4e6fda88db08\" (UID: \"2594a86c-caf2-4874-a415-4e6fda88db08\") " Dec 11 09:00:15 crc kubenswrapper[4629]: I1211 09:00:15.942193 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2594a86c-caf2-4874-a415-4e6fda88db08-kube-api-access-d8jnv" (OuterVolumeSpecName: "kube-api-access-d8jnv") pod "2594a86c-caf2-4874-a415-4e6fda88db08" (UID: "2594a86c-caf2-4874-a415-4e6fda88db08"). InnerVolumeSpecName "kube-api-access-d8jnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.039038 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8jnv\" (UniqueName: \"kubernetes.io/projected/2594a86c-caf2-4874-a415-4e6fda88db08-kube-api-access-d8jnv\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.209515 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5290303-73ff-47c7-8b4e-6518841d75ec" path="/var/lib/kubelet/pods/a5290303-73ff-47c7-8b4e-6518841d75ec/volumes" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.626649 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2594a86c-caf2-4874-a415-4e6fda88db08","Type":"ContainerDied","Data":"a8a55aaa6287412e36c6e0c6b921d76515e9b2c68eab039ac9479d99d9f4ac3d"} Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.627122 4629 scope.go:117] "RemoveContainer" containerID="1cd05b725be1d7583bf973ed8dba0ef5f55f3f153bef4abf0bc35e8be720622c" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.626924 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.658908 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.683477 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.719298 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 09:00:16 crc kubenswrapper[4629]: E1211 09:00:16.719717 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5290303-73ff-47c7-8b4e-6518841d75ec" containerName="dnsmasq-dns" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.719735 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5290303-73ff-47c7-8b4e-6518841d75ec" containerName="dnsmasq-dns" Dec 11 09:00:16 crc kubenswrapper[4629]: E1211 09:00:16.719758 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5290303-73ff-47c7-8b4e-6518841d75ec" containerName="init" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.719765 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5290303-73ff-47c7-8b4e-6518841d75ec" containerName="init" Dec 11 09:00:16 crc kubenswrapper[4629]: E1211 09:00:16.719796 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2594a86c-caf2-4874-a415-4e6fda88db08" containerName="kube-state-metrics" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.719802 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="2594a86c-caf2-4874-a415-4e6fda88db08" containerName="kube-state-metrics" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.720035 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5290303-73ff-47c7-8b4e-6518841d75ec" containerName="dnsmasq-dns" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.720054 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="2594a86c-caf2-4874-a415-4e6fda88db08" containerName="kube-state-metrics" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.720809 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.726241 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.726308 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.731292 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.755108 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/005eb316-fed1-4771-9934-3d86cc6d02d3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"005eb316-fed1-4771-9934-3d86cc6d02d3\") " pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.755243 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/005eb316-fed1-4771-9934-3d86cc6d02d3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"005eb316-fed1-4771-9934-3d86cc6d02d3\") " pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.755391 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/005eb316-fed1-4771-9934-3d86cc6d02d3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"005eb316-fed1-4771-9934-3d86cc6d02d3\") " pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.755427 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvdtx\" (UniqueName: \"kubernetes.io/projected/005eb316-fed1-4771-9934-3d86cc6d02d3-kube-api-access-gvdtx\") pod \"kube-state-metrics-0\" (UID: \"005eb316-fed1-4771-9934-3d86cc6d02d3\") " pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.859227 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/005eb316-fed1-4771-9934-3d86cc6d02d3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"005eb316-fed1-4771-9934-3d86cc6d02d3\") " pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.859353 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/005eb316-fed1-4771-9934-3d86cc6d02d3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"005eb316-fed1-4771-9934-3d86cc6d02d3\") " pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.859390 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvdtx\" (UniqueName: \"kubernetes.io/projected/005eb316-fed1-4771-9934-3d86cc6d02d3-kube-api-access-gvdtx\") pod \"kube-state-metrics-0\" (UID: \"005eb316-fed1-4771-9934-3d86cc6d02d3\") " pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.859460 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/005eb316-fed1-4771-9934-3d86cc6d02d3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"005eb316-fed1-4771-9934-3d86cc6d02d3\") " pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.865332 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/005eb316-fed1-4771-9934-3d86cc6d02d3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"005eb316-fed1-4771-9934-3d86cc6d02d3\") " pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.865678 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/005eb316-fed1-4771-9934-3d86cc6d02d3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"005eb316-fed1-4771-9934-3d86cc6d02d3\") " pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.870633 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/005eb316-fed1-4771-9934-3d86cc6d02d3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"005eb316-fed1-4771-9934-3d86cc6d02d3\") " pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.882831 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvdtx\" (UniqueName: \"kubernetes.io/projected/005eb316-fed1-4771-9934-3d86cc6d02d3-kube-api-access-gvdtx\") pod \"kube-state-metrics-0\" (UID: \"005eb316-fed1-4771-9934-3d86cc6d02d3\") " pod="openstack/kube-state-metrics-0" Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.991069 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.991393 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="ceilometer-central-agent" containerID="cri-o://65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630" gracePeriod=30 Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.991794 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="proxy-httpd" containerID="cri-o://7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9" gracePeriod=30 Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.991866 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="sg-core" containerID="cri-o://4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1" gracePeriod=30 Dec 11 09:00:16 crc kubenswrapper[4629]: I1211 09:00:16.991901 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="ceilometer-notification-agent" containerID="cri-o://cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205" gracePeriod=30 Dec 11 09:00:17 crc kubenswrapper[4629]: I1211 09:00:17.043198 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 09:00:17 crc kubenswrapper[4629]: I1211 09:00:17.590907 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 09:00:17 crc kubenswrapper[4629]: I1211 09:00:17.641044 4629 generic.go:334] "Generic (PLEG): container finished" podID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerID="7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9" exitCode=0 Dec 11 09:00:17 crc kubenswrapper[4629]: I1211 09:00:17.641081 4629 generic.go:334] "Generic (PLEG): container finished" podID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerID="4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1" exitCode=2 Dec 11 09:00:17 crc kubenswrapper[4629]: I1211 09:00:17.641092 4629 generic.go:334] "Generic (PLEG): container finished" podID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerID="65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630" exitCode=0 Dec 11 09:00:17 crc kubenswrapper[4629]: I1211 09:00:17.641150 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ad9572b-797d-4696-bd0a-2f6c388827ff","Type":"ContainerDied","Data":"7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9"} Dec 11 09:00:17 crc kubenswrapper[4629]: I1211 09:00:17.641186 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ad9572b-797d-4696-bd0a-2f6c388827ff","Type":"ContainerDied","Data":"4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1"} Dec 11 09:00:17 crc kubenswrapper[4629]: I1211 09:00:17.641199 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ad9572b-797d-4696-bd0a-2f6c388827ff","Type":"ContainerDied","Data":"65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630"} Dec 11 09:00:17 crc kubenswrapper[4629]: I1211 09:00:17.643375 4629 generic.go:334] "Generic (PLEG): container finished" podID="a5f5f734-35d7-49f9-a007-d1722d4d4673" containerID="a42cfdb3a956accc8e6e35bf29d2aa5681ece9adba1c92efc77474b2a3fb1f98" exitCode=0 Dec 11 09:00:17 crc kubenswrapper[4629]: I1211 09:00:17.643425 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vppnj" event={"ID":"a5f5f734-35d7-49f9-a007-d1722d4d4673","Type":"ContainerDied","Data":"a42cfdb3a956accc8e6e35bf29d2aa5681ece9adba1c92efc77474b2a3fb1f98"} Dec 11 09:00:17 crc kubenswrapper[4629]: I1211 09:00:17.646095 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"005eb316-fed1-4771-9934-3d86cc6d02d3","Type":"ContainerStarted","Data":"14d601964e347f211a7edc3dcb716b2dd6446deb36b6262b9ca39b9d59d27738"} Dec 11 09:00:18 crc kubenswrapper[4629]: I1211 09:00:18.211373 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2594a86c-caf2-4874-a415-4e6fda88db08" path="/var/lib/kubelet/pods/2594a86c-caf2-4874-a415-4e6fda88db08/volumes" Dec 11 09:00:18 crc kubenswrapper[4629]: I1211 09:00:18.656167 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"005eb316-fed1-4771-9934-3d86cc6d02d3","Type":"ContainerStarted","Data":"a8da6e2c05b1e2666c8ba4a891d706e757f9d19c51f16f2881d141d9c117bab1"} Dec 11 09:00:18 crc kubenswrapper[4629]: I1211 09:00:18.656346 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 11 09:00:18 crc kubenswrapper[4629]: I1211 09:00:18.657666 4629 generic.go:334] "Generic (PLEG): container finished" podID="8f83bf9d-806c-4310-9315-2bcbd529a6d1" containerID="9f9ffba2e0fd7c28249813b71c9e26135de39e68728e184c968d7475561fa5ba" exitCode=0 Dec 11 09:00:18 crc kubenswrapper[4629]: I1211 09:00:18.657861 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-65bhx" event={"ID":"8f83bf9d-806c-4310-9315-2bcbd529a6d1","Type":"ContainerDied","Data":"9f9ffba2e0fd7c28249813b71c9e26135de39e68728e184c968d7475561fa5ba"} Dec 11 09:00:18 crc kubenswrapper[4629]: I1211 09:00:18.687209 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.297742722 podStartE2EDuration="2.687189159s" podCreationTimestamp="2025-12-11 09:00:16 +0000 UTC" firstStartedPulling="2025-12-11 09:00:17.599296023 +0000 UTC m=+1405.664711631" lastFinishedPulling="2025-12-11 09:00:17.98874246 +0000 UTC m=+1406.054158068" observedRunningTime="2025-12-11 09:00:18.67862696 +0000 UTC m=+1406.744042578" watchObservedRunningTime="2025-12-11 09:00:18.687189159 +0000 UTC m=+1406.752604767" Dec 11 09:00:18 crc kubenswrapper[4629]: I1211 09:00:18.928288 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 09:00:18 crc kubenswrapper[4629]: I1211 09:00:18.928748 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.044565 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.111158 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-combined-ca-bundle\") pod \"a5f5f734-35d7-49f9-a007-d1722d4d4673\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.111228 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4hrq\" (UniqueName: \"kubernetes.io/projected/a5f5f734-35d7-49f9-a007-d1722d4d4673-kube-api-access-g4hrq\") pod \"a5f5f734-35d7-49f9-a007-d1722d4d4673\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.111281 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-scripts\") pod \"a5f5f734-35d7-49f9-a007-d1722d4d4673\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.111413 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-config-data\") pod \"a5f5f734-35d7-49f9-a007-d1722d4d4673\" (UID: \"a5f5f734-35d7-49f9-a007-d1722d4d4673\") " Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.119748 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5f5f734-35d7-49f9-a007-d1722d4d4673-kube-api-access-g4hrq" (OuterVolumeSpecName: "kube-api-access-g4hrq") pod "a5f5f734-35d7-49f9-a007-d1722d4d4673" (UID: "a5f5f734-35d7-49f9-a007-d1722d4d4673"). InnerVolumeSpecName "kube-api-access-g4hrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.119987 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-scripts" (OuterVolumeSpecName: "scripts") pod "a5f5f734-35d7-49f9-a007-d1722d4d4673" (UID: "a5f5f734-35d7-49f9-a007-d1722d4d4673"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.149605 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5f5f734-35d7-49f9-a007-d1722d4d4673" (UID: "a5f5f734-35d7-49f9-a007-d1722d4d4673"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.150769 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-config-data" (OuterVolumeSpecName: "config-data") pod "a5f5f734-35d7-49f9-a007-d1722d4d4673" (UID: "a5f5f734-35d7-49f9-a007-d1722d4d4673"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.213930 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.214161 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.214278 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4hrq\" (UniqueName: \"kubernetes.io/projected/a5f5f734-35d7-49f9-a007-d1722d4d4673-kube-api-access-g4hrq\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.214377 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f5f734-35d7-49f9-a007-d1722d4d4673-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.668020 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vppnj" event={"ID":"a5f5f734-35d7-49f9-a007-d1722d4d4673","Type":"ContainerDied","Data":"ef33646d470b46b86370961feef7dede3ee7fbdff54fa7b41260db0729f3c949"} Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.668184 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vppnj" Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.670055 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef33646d470b46b86370961feef7dede3ee7fbdff54fa7b41260db0729f3c949" Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.845482 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.846296 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="64a3e1bc-8fa8-4493-ac7c-865e51169162" containerName="nova-api-api" containerID="cri-o://ba34b4bae6ba99fb9c0b068508d3ec43556b0216a99b57085667ebc210a1e399" gracePeriod=30 Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.846237 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="64a3e1bc-8fa8-4493-ac7c-865e51169162" containerName="nova-api-log" containerID="cri-o://afa498f6abcebe5d64954bd9a552c17c9ff1a92839177a5c95be94fecc0e3ef9" gracePeriod=30 Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.879977 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.880188 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a8336f4a-02d9-4423-b40f-231ee4781b89" containerName="nova-scheduler-scheduler" containerID="cri-o://345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9" gracePeriod=30 Dec 11 09:00:19 crc kubenswrapper[4629]: I1211 09:00:19.911111 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.139950 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.332723 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-combined-ca-bundle\") pod \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.332802 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-scripts\") pod \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.332836 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-config-data\") pod \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.332919 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhw6z\" (UniqueName: \"kubernetes.io/projected/8f83bf9d-806c-4310-9315-2bcbd529a6d1-kube-api-access-xhw6z\") pod \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\" (UID: \"8f83bf9d-806c-4310-9315-2bcbd529a6d1\") " Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.342971 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-scripts" (OuterVolumeSpecName: "scripts") pod "8f83bf9d-806c-4310-9315-2bcbd529a6d1" (UID: "8f83bf9d-806c-4310-9315-2bcbd529a6d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.343113 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f83bf9d-806c-4310-9315-2bcbd529a6d1-kube-api-access-xhw6z" (OuterVolumeSpecName: "kube-api-access-xhw6z") pod "8f83bf9d-806c-4310-9315-2bcbd529a6d1" (UID: "8f83bf9d-806c-4310-9315-2bcbd529a6d1"). InnerVolumeSpecName "kube-api-access-xhw6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.367412 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-config-data" (OuterVolumeSpecName: "config-data") pod "8f83bf9d-806c-4310-9315-2bcbd529a6d1" (UID: "8f83bf9d-806c-4310-9315-2bcbd529a6d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.390398 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f83bf9d-806c-4310-9315-2bcbd529a6d1" (UID: "8f83bf9d-806c-4310-9315-2bcbd529a6d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.435618 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.435662 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.435676 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f83bf9d-806c-4310-9315-2bcbd529a6d1-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.435688 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhw6z\" (UniqueName: \"kubernetes.io/projected/8f83bf9d-806c-4310-9315-2bcbd529a6d1-kube-api-access-xhw6z\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.485919 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.485978 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.486021 4629 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.486756 4629 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f34698209a804ac13c8c04a3199ac3381f69055cef9be6fa18f868fa21af6591"} pod="openshift-machine-config-operator/machine-config-daemon-wbld9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.486813 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" containerID="cri-o://f34698209a804ac13c8c04a3199ac3381f69055cef9be6fa18f868fa21af6591" gracePeriod=600 Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.679913 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-65bhx" event={"ID":"8f83bf9d-806c-4310-9315-2bcbd529a6d1","Type":"ContainerDied","Data":"6c90fcb38758fe12e6b0f10c34bdf81935e886ea357bb959da463f5e1ad9401b"} Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.680280 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c90fcb38758fe12e6b0f10c34bdf81935e886ea357bb959da463f5e1ad9401b" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.680964 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-65bhx" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.683206 4629 generic.go:334] "Generic (PLEG): container finished" podID="64a3e1bc-8fa8-4493-ac7c-865e51169162" containerID="afa498f6abcebe5d64954bd9a552c17c9ff1a92839177a5c95be94fecc0e3ef9" exitCode=143 Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.683267 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64a3e1bc-8fa8-4493-ac7c-865e51169162","Type":"ContainerDied","Data":"afa498f6abcebe5d64954bd9a552c17c9ff1a92839177a5c95be94fecc0e3ef9"} Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.687559 4629 generic.go:334] "Generic (PLEG): container finished" podID="847b50b9-0093-404f-b0dd-63c814d8986b" containerID="f34698209a804ac13c8c04a3199ac3381f69055cef9be6fa18f868fa21af6591" exitCode=0 Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.687756 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e708d4fe-3394-42cf-a1b8-e93804c8fd6d" containerName="nova-metadata-log" containerID="cri-o://b1392b230a8cb39edecd1065c18563af1246b312d49cea2b5d533cc90bec0a0e" gracePeriod=30 Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.687844 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerDied","Data":"f34698209a804ac13c8c04a3199ac3381f69055cef9be6fa18f868fa21af6591"} Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.687894 4629 scope.go:117] "RemoveContainer" containerID="921d43c980dda9b89debb49d73c6ac3dda5f931588d36bb68efdef51a59b46b8" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.688267 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e708d4fe-3394-42cf-a1b8-e93804c8fd6d" containerName="nova-metadata-metadata" containerID="cri-o://cc07115752980f532522f711989a142b1be931d1c2dc347dd40a23d2e15d5880" gracePeriod=30 Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.790332 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 11 09:00:20 crc kubenswrapper[4629]: E1211 09:00:20.790773 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f83bf9d-806c-4310-9315-2bcbd529a6d1" containerName="nova-cell1-conductor-db-sync" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.790799 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f83bf9d-806c-4310-9315-2bcbd529a6d1" containerName="nova-cell1-conductor-db-sync" Dec 11 09:00:20 crc kubenswrapper[4629]: E1211 09:00:20.790819 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f5f734-35d7-49f9-a007-d1722d4d4673" containerName="nova-manage" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.790832 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f5f734-35d7-49f9-a007-d1722d4d4673" containerName="nova-manage" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.791026 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f5f734-35d7-49f9-a007-d1722d4d4673" containerName="nova-manage" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.791066 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f83bf9d-806c-4310-9315-2bcbd529a6d1" containerName="nova-cell1-conductor-db-sync" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.791765 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.796475 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.815343 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.957996 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq67h\" (UniqueName: \"kubernetes.io/projected/e3ea5104-e7b6-41fc-a04a-e1292739c878-kube-api-access-fq67h\") pod \"nova-cell1-conductor-0\" (UID: \"e3ea5104-e7b6-41fc-a04a-e1292739c878\") " pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.958093 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ea5104-e7b6-41fc-a04a-e1292739c878-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e3ea5104-e7b6-41fc-a04a-e1292739c878\") " pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:20 crc kubenswrapper[4629]: I1211 09:00:20.958143 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ea5104-e7b6-41fc-a04a-e1292739c878-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e3ea5104-e7b6-41fc-a04a-e1292739c878\") " pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:21 crc kubenswrapper[4629]: I1211 09:00:21.059342 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq67h\" (UniqueName: \"kubernetes.io/projected/e3ea5104-e7b6-41fc-a04a-e1292739c878-kube-api-access-fq67h\") pod \"nova-cell1-conductor-0\" (UID: \"e3ea5104-e7b6-41fc-a04a-e1292739c878\") " pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:21 crc kubenswrapper[4629]: I1211 09:00:21.059430 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ea5104-e7b6-41fc-a04a-e1292739c878-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e3ea5104-e7b6-41fc-a04a-e1292739c878\") " pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:21 crc kubenswrapper[4629]: I1211 09:00:21.059482 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ea5104-e7b6-41fc-a04a-e1292739c878-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e3ea5104-e7b6-41fc-a04a-e1292739c878\") " pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:21 crc kubenswrapper[4629]: I1211 09:00:21.066957 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ea5104-e7b6-41fc-a04a-e1292739c878-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e3ea5104-e7b6-41fc-a04a-e1292739c878\") " pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:21 crc kubenswrapper[4629]: I1211 09:00:21.067450 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ea5104-e7b6-41fc-a04a-e1292739c878-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e3ea5104-e7b6-41fc-a04a-e1292739c878\") " pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:21 crc kubenswrapper[4629]: I1211 09:00:21.080568 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq67h\" (UniqueName: \"kubernetes.io/projected/e3ea5104-e7b6-41fc-a04a-e1292739c878-kube-api-access-fq67h\") pod \"nova-cell1-conductor-0\" (UID: \"e3ea5104-e7b6-41fc-a04a-e1292739c878\") " pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:22 crc kubenswrapper[4629]: I1211 09:00:22.012669 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:22 crc kubenswrapper[4629]: I1211 09:00:22.531928 4629 generic.go:334] "Generic (PLEG): container finished" podID="e708d4fe-3394-42cf-a1b8-e93804c8fd6d" containerID="b1392b230a8cb39edecd1065c18563af1246b312d49cea2b5d533cc90bec0a0e" exitCode=143 Dec 11 09:00:22 crc kubenswrapper[4629]: I1211 09:00:22.533905 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e708d4fe-3394-42cf-a1b8-e93804c8fd6d","Type":"ContainerDied","Data":"b1392b230a8cb39edecd1065c18563af1246b312d49cea2b5d533cc90bec0a0e"} Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.055814 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.418546 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.509455 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz5j2\" (UniqueName: \"kubernetes.io/projected/8ad9572b-797d-4696-bd0a-2f6c388827ff-kube-api-access-nz5j2\") pod \"8ad9572b-797d-4696-bd0a-2f6c388827ff\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.509823 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-config-data\") pod \"8ad9572b-797d-4696-bd0a-2f6c388827ff\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.509956 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ad9572b-797d-4696-bd0a-2f6c388827ff-log-httpd\") pod \"8ad9572b-797d-4696-bd0a-2f6c388827ff\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.510012 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ad9572b-797d-4696-bd0a-2f6c388827ff-run-httpd\") pod \"8ad9572b-797d-4696-bd0a-2f6c388827ff\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.510040 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-scripts\") pod \"8ad9572b-797d-4696-bd0a-2f6c388827ff\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.510111 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-sg-core-conf-yaml\") pod \"8ad9572b-797d-4696-bd0a-2f6c388827ff\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.510169 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-combined-ca-bundle\") pod \"8ad9572b-797d-4696-bd0a-2f6c388827ff\" (UID: \"8ad9572b-797d-4696-bd0a-2f6c388827ff\") " Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.611620 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ad9572b-797d-4696-bd0a-2f6c388827ff-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8ad9572b-797d-4696-bd0a-2f6c388827ff" (UID: "8ad9572b-797d-4696-bd0a-2f6c388827ff"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.614107 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad9572b-797d-4696-bd0a-2f6c388827ff-kube-api-access-nz5j2" (OuterVolumeSpecName: "kube-api-access-nz5j2") pod "8ad9572b-797d-4696-bd0a-2f6c388827ff" (UID: "8ad9572b-797d-4696-bd0a-2f6c388827ff"). InnerVolumeSpecName "kube-api-access-nz5j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.617162 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ad9572b-797d-4696-bd0a-2f6c388827ff-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8ad9572b-797d-4696-bd0a-2f6c388827ff" (UID: "8ad9572b-797d-4696-bd0a-2f6c388827ff"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.624607 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz5j2\" (UniqueName: \"kubernetes.io/projected/8ad9572b-797d-4696-bd0a-2f6c388827ff-kube-api-access-nz5j2\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.624655 4629 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ad9572b-797d-4696-bd0a-2f6c388827ff-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.624666 4629 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ad9572b-797d-4696-bd0a-2f6c388827ff-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.626101 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-scripts" (OuterVolumeSpecName: "scripts") pod "8ad9572b-797d-4696-bd0a-2f6c388827ff" (UID: "8ad9572b-797d-4696-bd0a-2f6c388827ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.645261 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e3ea5104-e7b6-41fc-a04a-e1292739c878","Type":"ContainerStarted","Data":"9e284c77798998eda2951f053a988e9c85ecfafebf2901ebb2d4f35f2fd73a8f"} Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.645327 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e3ea5104-e7b6-41fc-a04a-e1292739c878","Type":"ContainerStarted","Data":"3353fb5ee6a141a885e0a01d6e5af6c8cd369cb5f731e9a151c6e1459e5e80c3"} Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.646969 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.648593 4629 generic.go:334] "Generic (PLEG): container finished" podID="64a3e1bc-8fa8-4493-ac7c-865e51169162" containerID="ba34b4bae6ba99fb9c0b068508d3ec43556b0216a99b57085667ebc210a1e399" exitCode=0 Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.648717 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64a3e1bc-8fa8-4493-ac7c-865e51169162","Type":"ContainerDied","Data":"ba34b4bae6ba99fb9c0b068508d3ec43556b0216a99b57085667ebc210a1e399"} Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.653744 4629 generic.go:334] "Generic (PLEG): container finished" podID="e708d4fe-3394-42cf-a1b8-e93804c8fd6d" containerID="cc07115752980f532522f711989a142b1be931d1c2dc347dd40a23d2e15d5880" exitCode=0 Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.653804 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e708d4fe-3394-42cf-a1b8-e93804c8fd6d","Type":"ContainerDied","Data":"cc07115752980f532522f711989a142b1be931d1c2dc347dd40a23d2e15d5880"} Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.658374 4629 generic.go:334] "Generic (PLEG): container finished" podID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerID="cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205" exitCode=0 Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.658450 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ad9572b-797d-4696-bd0a-2f6c388827ff","Type":"ContainerDied","Data":"cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205"} Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.658484 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ad9572b-797d-4696-bd0a-2f6c388827ff","Type":"ContainerDied","Data":"4caf9f8da6169afa1b33348aad1e7e44acce078db64489f1a9dd11513b4b9088"} Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.658505 4629 scope.go:117] "RemoveContainer" containerID="7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.658650 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.669184 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f"} Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.709805 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.709781468 podStartE2EDuration="3.709781468s" podCreationTimestamp="2025-12-11 09:00:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:23.682073558 +0000 UTC m=+1411.747489176" watchObservedRunningTime="2025-12-11 09:00:23.709781468 +0000 UTC m=+1411.775197076" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.727224 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.739055 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8ad9572b-797d-4696-bd0a-2f6c388827ff" (UID: "8ad9572b-797d-4696-bd0a-2f6c388827ff"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.765287 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-config-data" (OuterVolumeSpecName: "config-data") pod "8ad9572b-797d-4696-bd0a-2f6c388827ff" (UID: "8ad9572b-797d-4696-bd0a-2f6c388827ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.775043 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ad9572b-797d-4696-bd0a-2f6c388827ff" (UID: "8ad9572b-797d-4696-bd0a-2f6c388827ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.840159 4629 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.840188 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.840198 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad9572b-797d-4696-bd0a-2f6c388827ff-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.921142 4629 scope.go:117] "RemoveContainer" containerID="4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1" Dec 11 09:00:23 crc kubenswrapper[4629]: I1211 09:00:23.970016 4629 scope.go:117] "RemoveContainer" containerID="cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.005288 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.029113 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.044096 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.045045 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64a3e1bc-8fa8-4493-ac7c-865e51169162-logs\") pod \"64a3e1bc-8fa8-4493-ac7c-865e51169162\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.051801 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64a3e1bc-8fa8-4493-ac7c-865e51169162-combined-ca-bundle\") pod \"64a3e1bc-8fa8-4493-ac7c-865e51169162\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.051875 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpv55\" (UniqueName: \"kubernetes.io/projected/64a3e1bc-8fa8-4493-ac7c-865e51169162-kube-api-access-kpv55\") pod \"64a3e1bc-8fa8-4493-ac7c-865e51169162\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.051907 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64a3e1bc-8fa8-4493-ac7c-865e51169162-config-data\") pod \"64a3e1bc-8fa8-4493-ac7c-865e51169162\" (UID: \"64a3e1bc-8fa8-4493-ac7c-865e51169162\") " Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.097024 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.099896 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="ceilometer-notification-agent" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.099926 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="ceilometer-notification-agent" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.099940 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="ceilometer-central-agent" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.099947 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="ceilometer-central-agent" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.099957 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a3e1bc-8fa8-4493-ac7c-865e51169162" containerName="nova-api-api" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.099966 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a3e1bc-8fa8-4493-ac7c-865e51169162" containerName="nova-api-api" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.099993 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a3e1bc-8fa8-4493-ac7c-865e51169162" containerName="nova-api-log" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.100002 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a3e1bc-8fa8-4493-ac7c-865e51169162" containerName="nova-api-log" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.100016 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="proxy-httpd" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.100023 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="proxy-httpd" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.100040 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="sg-core" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.100046 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="sg-core" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.100287 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="ceilometer-central-agent" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.100303 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="sg-core" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.100317 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="64a3e1bc-8fa8-4493-ac7c-865e51169162" containerName="nova-api-api" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.100333 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="proxy-httpd" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.100351 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" containerName="ceilometer-notification-agent" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.100363 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="64a3e1bc-8fa8-4493-ac7c-865e51169162" containerName="nova-api-log" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.102796 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.109507 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.110500 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.112177 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.124876 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.131478 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64a3e1bc-8fa8-4493-ac7c-865e51169162-logs" (OuterVolumeSpecName: "logs") pod "64a3e1bc-8fa8-4493-ac7c-865e51169162" (UID: "64a3e1bc-8fa8-4493-ac7c-865e51169162"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.136427 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64a3e1bc-8fa8-4493-ac7c-865e51169162-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64a3e1bc-8fa8-4493-ac7c-865e51169162" (UID: "64a3e1bc-8fa8-4493-ac7c-865e51169162"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.138035 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64a3e1bc-8fa8-4493-ac7c-865e51169162-config-data" (OuterVolumeSpecName: "config-data") pod "64a3e1bc-8fa8-4493-ac7c-865e51169162" (UID: "64a3e1bc-8fa8-4493-ac7c-865e51169162"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.143047 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64a3e1bc-8fa8-4493-ac7c-865e51169162-kube-api-access-kpv55" (OuterVolumeSpecName: "kube-api-access-kpv55") pod "64a3e1bc-8fa8-4493-ac7c-865e51169162" (UID: "64a3e1bc-8fa8-4493-ac7c-865e51169162"). InnerVolumeSpecName "kube-api-access-kpv55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.154039 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-config-data\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.154097 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.154199 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.154228 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57d76178-7b70-4fc6-a865-579eed1fefe6-run-httpd\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.154258 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57d76178-7b70-4fc6-a865-579eed1fefe6-log-httpd\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.154284 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-scripts\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.154388 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.154430 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lss5l\" (UniqueName: \"kubernetes.io/projected/57d76178-7b70-4fc6-a865-579eed1fefe6-kube-api-access-lss5l\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.154513 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64a3e1bc-8fa8-4493-ac7c-865e51169162-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.154527 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpv55\" (UniqueName: \"kubernetes.io/projected/64a3e1bc-8fa8-4493-ac7c-865e51169162-kube-api-access-kpv55\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.154539 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64a3e1bc-8fa8-4493-ac7c-865e51169162-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.154548 4629 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64a3e1bc-8fa8-4493-ac7c-865e51169162-logs\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.212514 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad9572b-797d-4696-bd0a-2f6c388827ff" path="/var/lib/kubelet/pods/8ad9572b-797d-4696-bd0a-2f6c388827ff/volumes" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.225000 4629 scope.go:117] "RemoveContainer" containerID="65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.225232 4629 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9 is running failed: container process not found" containerID="345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.225519 4629 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9 is running failed: container process not found" containerID="345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.226161 4629 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9 is running failed: container process not found" containerID="345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.226198 4629 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a8336f4a-02d9-4423-b40f-231ee4781b89" containerName="nova-scheduler-scheduler" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.255693 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-config-data\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.264146 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.255747 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.265253 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.265297 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57d76178-7b70-4fc6-a865-579eed1fefe6-run-httpd\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.265347 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57d76178-7b70-4fc6-a865-579eed1fefe6-log-httpd\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.265383 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-scripts\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.265477 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.265533 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lss5l\" (UniqueName: \"kubernetes.io/projected/57d76178-7b70-4fc6-a865-579eed1fefe6-kube-api-access-lss5l\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.266589 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57d76178-7b70-4fc6-a865-579eed1fefe6-log-httpd\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.267142 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57d76178-7b70-4fc6-a865-579eed1fefe6-run-httpd\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.269255 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.272602 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-scripts\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.277400 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.277723 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.280749 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-config-data\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.301124 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lss5l\" (UniqueName: \"kubernetes.io/projected/57d76178-7b70-4fc6-a865-579eed1fefe6-kube-api-access-lss5l\") pod \"ceilometer-0\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.323506 4629 scope.go:117] "RemoveContainer" containerID="7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.324099 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9\": container with ID starting with 7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9 not found: ID does not exist" containerID="7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.324131 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9"} err="failed to get container status \"7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9\": rpc error: code = NotFound desc = could not find container \"7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9\": container with ID starting with 7eb703ff3ef170654476ee5c8f31bef142a6d65ea4ec98093d8b3c3e8a4284a9 not found: ID does not exist" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.324158 4629 scope.go:117] "RemoveContainer" containerID="4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.333918 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1\": container with ID starting with 4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1 not found: ID does not exist" containerID="4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.334004 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1"} err="failed to get container status \"4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1\": rpc error: code = NotFound desc = could not find container \"4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1\": container with ID starting with 4467e793c43625374d496e4a253ce7d3bba4f1c749d171d6e311a3b8635c79f1 not found: ID does not exist" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.334038 4629 scope.go:117] "RemoveContainer" containerID="cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.334674 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205\": container with ID starting with cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205 not found: ID does not exist" containerID="cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.334726 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205"} err="failed to get container status \"cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205\": rpc error: code = NotFound desc = could not find container \"cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205\": container with ID starting with cabd1c9a7ef85ac4db452df692a031b226d82b2b148e7525496fc14fb03f3205 not found: ID does not exist" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.334748 4629 scope.go:117] "RemoveContainer" containerID="65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.335099 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630\": container with ID starting with 65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630 not found: ID does not exist" containerID="65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.335127 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630"} err="failed to get container status \"65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630\": rpc error: code = NotFound desc = could not find container \"65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630\": container with ID starting with 65ff29138d9c5e211ececc483bfcde0bda76d601fe1c8431e62617f8404be630 not found: ID does not exist" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.368568 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-nova-metadata-tls-certs\") pod \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.368614 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqbqp\" (UniqueName: \"kubernetes.io/projected/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-kube-api-access-kqbqp\") pod \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.368696 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-config-data\") pod \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.368819 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-logs\") pod \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.368909 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-combined-ca-bundle\") pod \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\" (UID: \"e708d4fe-3394-42cf-a1b8-e93804c8fd6d\") " Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.385570 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-logs" (OuterVolumeSpecName: "logs") pod "e708d4fe-3394-42cf-a1b8-e93804c8fd6d" (UID: "e708d4fe-3394-42cf-a1b8-e93804c8fd6d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.392280 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.393064 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-kube-api-access-kqbqp" (OuterVolumeSpecName: "kube-api-access-kqbqp") pod "e708d4fe-3394-42cf-a1b8-e93804c8fd6d" (UID: "e708d4fe-3394-42cf-a1b8-e93804c8fd6d"). InnerVolumeSpecName "kube-api-access-kqbqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.404748 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e708d4fe-3394-42cf-a1b8-e93804c8fd6d" (UID: "e708d4fe-3394-42cf-a1b8-e93804c8fd6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.405438 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-config-data" (OuterVolumeSpecName: "config-data") pod "e708d4fe-3394-42cf-a1b8-e93804c8fd6d" (UID: "e708d4fe-3394-42cf-a1b8-e93804c8fd6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.431136 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.470264 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8336f4a-02d9-4423-b40f-231ee4781b89-combined-ca-bundle\") pod \"a8336f4a-02d9-4423-b40f-231ee4781b89\" (UID: \"a8336f4a-02d9-4423-b40f-231ee4781b89\") " Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.470581 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l7x6\" (UniqueName: \"kubernetes.io/projected/a8336f4a-02d9-4423-b40f-231ee4781b89-kube-api-access-2l7x6\") pod \"a8336f4a-02d9-4423-b40f-231ee4781b89\" (UID: \"a8336f4a-02d9-4423-b40f-231ee4781b89\") " Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.470733 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8336f4a-02d9-4423-b40f-231ee4781b89-config-data\") pod \"a8336f4a-02d9-4423-b40f-231ee4781b89\" (UID: \"a8336f4a-02d9-4423-b40f-231ee4781b89\") " Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.471278 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.471291 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqbqp\" (UniqueName: \"kubernetes.io/projected/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-kube-api-access-kqbqp\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.471304 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.471332 4629 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-logs\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.473302 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "e708d4fe-3394-42cf-a1b8-e93804c8fd6d" (UID: "e708d4fe-3394-42cf-a1b8-e93804c8fd6d"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.478033 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8336f4a-02d9-4423-b40f-231ee4781b89-kube-api-access-2l7x6" (OuterVolumeSpecName: "kube-api-access-2l7x6") pod "a8336f4a-02d9-4423-b40f-231ee4781b89" (UID: "a8336f4a-02d9-4423-b40f-231ee4781b89"). InnerVolumeSpecName "kube-api-access-2l7x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.497556 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8336f4a-02d9-4423-b40f-231ee4781b89-config-data" (OuterVolumeSpecName: "config-data") pod "a8336f4a-02d9-4423-b40f-231ee4781b89" (UID: "a8336f4a-02d9-4423-b40f-231ee4781b89"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.520784 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8336f4a-02d9-4423-b40f-231ee4781b89-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8336f4a-02d9-4423-b40f-231ee4781b89" (UID: "a8336f4a-02d9-4423-b40f-231ee4781b89"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.588244 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8336f4a-02d9-4423-b40f-231ee4781b89-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.588602 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8336f4a-02d9-4423-b40f-231ee4781b89-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.588621 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l7x6\" (UniqueName: \"kubernetes.io/projected/a8336f4a-02d9-4423-b40f-231ee4781b89-kube-api-access-2l7x6\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.588635 4629 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e708d4fe-3394-42cf-a1b8-e93804c8fd6d-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.680976 4629 generic.go:334] "Generic (PLEG): container finished" podID="a8336f4a-02d9-4423-b40f-231ee4781b89" containerID="345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9" exitCode=0 Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.681105 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.681163 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a8336f4a-02d9-4423-b40f-231ee4781b89","Type":"ContainerDied","Data":"345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9"} Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.681249 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a8336f4a-02d9-4423-b40f-231ee4781b89","Type":"ContainerDied","Data":"0d9119215206da99d1199e4762f08f333997526d94d0d017a78fd4655afe5b20"} Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.681272 4629 scope.go:117] "RemoveContainer" containerID="345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.691373 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64a3e1bc-8fa8-4493-ac7c-865e51169162","Type":"ContainerDied","Data":"5f09c59420a9db9f9a4d10fcd99c4913288c4bf0a0920fb4c7ba739336791714"} Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.691487 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.748152 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e708d4fe-3394-42cf-a1b8-e93804c8fd6d","Type":"ContainerDied","Data":"7eea11652b7ef26d4cdec084beea387d04cda184c63a6ca0bb26cffbe7820156"} Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.748280 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.777566 4629 scope.go:117] "RemoveContainer" containerID="345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.778000 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9\": container with ID starting with 345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9 not found: ID does not exist" containerID="345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.778031 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9"} err="failed to get container status \"345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9\": rpc error: code = NotFound desc = could not find container \"345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9\": container with ID starting with 345e7808a38e440f6a67d06733361bc86565995bda04d2afb6ddac4e6696bbc9 not found: ID does not exist" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.778056 4629 scope.go:117] "RemoveContainer" containerID="ba34b4bae6ba99fb9c0b068508d3ec43556b0216a99b57085667ebc210a1e399" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.819471 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.852329 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.853160 4629 scope.go:117] "RemoveContainer" containerID="afa498f6abcebe5d64954bd9a552c17c9ff1a92839177a5c95be94fecc0e3ef9" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.864516 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.890370 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.890869 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e708d4fe-3394-42cf-a1b8-e93804c8fd6d" containerName="nova-metadata-metadata" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.890892 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="e708d4fe-3394-42cf-a1b8-e93804c8fd6d" containerName="nova-metadata-metadata" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.890919 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8336f4a-02d9-4423-b40f-231ee4781b89" containerName="nova-scheduler-scheduler" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.890927 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8336f4a-02d9-4423-b40f-231ee4781b89" containerName="nova-scheduler-scheduler" Dec 11 09:00:24 crc kubenswrapper[4629]: E1211 09:00:24.890946 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e708d4fe-3394-42cf-a1b8-e93804c8fd6d" containerName="nova-metadata-log" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.890954 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="e708d4fe-3394-42cf-a1b8-e93804c8fd6d" containerName="nova-metadata-log" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.891164 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8336f4a-02d9-4423-b40f-231ee4781b89" containerName="nova-scheduler-scheduler" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.891178 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="e708d4fe-3394-42cf-a1b8-e93804c8fd6d" containerName="nova-metadata-metadata" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.891186 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="e708d4fe-3394-42cf-a1b8-e93804c8fd6d" containerName="nova-metadata-log" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.892388 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.894800 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.908479 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.922572 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.933329 4629 scope.go:117] "RemoveContainer" containerID="cc07115752980f532522f711989a142b1be931d1c2dc347dd40a23d2e15d5880" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.937911 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.939695 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.944471 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.960469 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:24 crc kubenswrapper[4629]: I1211 09:00:24.976395 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.000915 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.012227 4629 scope.go:117] "RemoveContainer" containerID="b1392b230a8cb39edecd1065c18563af1246b312d49cea2b5d533cc90bec0a0e" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.012455 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.014087 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-config-data\") pod \"nova-scheduler-0\" (UID: \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.014131 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-logs\") pod \"nova-api-0\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.014228 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc22j\" (UniqueName: \"kubernetes.io/projected/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-kube-api-access-zc22j\") pod \"nova-scheduler-0\" (UID: \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.014344 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzmxn\" (UniqueName: \"kubernetes.io/projected/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-kube-api-access-fzmxn\") pod \"nova-api-0\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.014403 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-config-data\") pod \"nova-api-0\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.014483 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.014514 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.015202 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.019720 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.019914 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.028983 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.080737 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.115788 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.115950 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzmxn\" (UniqueName: \"kubernetes.io/projected/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-kube-api-access-fzmxn\") pod \"nova-api-0\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.115986 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.116025 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-config-data\") pod \"nova-api-0\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.116081 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lv5m\" (UniqueName: \"kubernetes.io/projected/7be44442-eb08-4028-87d9-96decf3215b9-kube-api-access-7lv5m\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.116104 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-config-data\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.116143 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.116172 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.116192 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7be44442-eb08-4028-87d9-96decf3215b9-logs\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.116271 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-config-data\") pod \"nova-scheduler-0\" (UID: \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.116309 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-logs\") pod \"nova-api-0\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.117379 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-logs\") pod \"nova-api-0\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.117843 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc22j\" (UniqueName: \"kubernetes.io/projected/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-kube-api-access-zc22j\") pod \"nova-scheduler-0\" (UID: \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.126955 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-config-data\") pod \"nova-scheduler-0\" (UID: \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.135052 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.136048 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.143005 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-config-data\") pod \"nova-api-0\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.147799 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc22j\" (UniqueName: \"kubernetes.io/projected/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-kube-api-access-zc22j\") pod \"nova-scheduler-0\" (UID: \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\") " pod="openstack/nova-scheduler-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.169719 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzmxn\" (UniqueName: \"kubernetes.io/projected/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-kube-api-access-fzmxn\") pod \"nova-api-0\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.219404 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.219462 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.219512 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lv5m\" (UniqueName: \"kubernetes.io/projected/7be44442-eb08-4028-87d9-96decf3215b9-kube-api-access-7lv5m\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.219571 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-config-data\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.219633 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7be44442-eb08-4028-87d9-96decf3215b9-logs\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.220428 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7be44442-eb08-4028-87d9-96decf3215b9-logs\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.222863 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.223937 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.224894 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-config-data\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.229926 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.268614 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.283497 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lv5m\" (UniqueName: \"kubernetes.io/projected/7be44442-eb08-4028-87d9-96decf3215b9-kube-api-access-7lv5m\") pod \"nova-metadata-0\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.384073 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.787506 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57d76178-7b70-4fc6-a865-579eed1fefe6","Type":"ContainerStarted","Data":"0fcd6c252e8de703ba3100cd6c40f0704e2e26161f2ade5ef6cede8e6e736a64"} Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.919019 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:00:25 crc kubenswrapper[4629]: W1211 09:00:25.933959 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4c82eb1_f999_45b7_82af_58e3b56a2e0a.slice/crio-62722dd553f0f93dc04a5f464d231d41522dd12e4c2765f260f0264f09bee538 WatchSource:0}: Error finding container 62722dd553f0f93dc04a5f464d231d41522dd12e4c2765f260f0264f09bee538: Status 404 returned error can't find the container with id 62722dd553f0f93dc04a5f464d231d41522dd12e4c2765f260f0264f09bee538 Dec 11 09:00:25 crc kubenswrapper[4629]: I1211 09:00:25.991621 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.060075 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.236139 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64a3e1bc-8fa8-4493-ac7c-865e51169162" path="/var/lib/kubelet/pods/64a3e1bc-8fa8-4493-ac7c-865e51169162/volumes" Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.237135 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8336f4a-02d9-4423-b40f-231ee4781b89" path="/var/lib/kubelet/pods/a8336f4a-02d9-4423-b40f-231ee4781b89/volumes" Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.237747 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e708d4fe-3394-42cf-a1b8-e93804c8fd6d" path="/var/lib/kubelet/pods/e708d4fe-3394-42cf-a1b8-e93804c8fd6d/volumes" Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.810132 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105","Type":"ContainerStarted","Data":"83cd8eeec150772c31861c882ccceed6aa5c21377536575900682c7209bea2eb"} Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.810484 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105","Type":"ContainerStarted","Data":"2d9b89d6eb5a1a43bb565dbbe1f5ce122c1bd5c131653d9eac3eeaf3c39f9d7a"} Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.810495 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105","Type":"ContainerStarted","Data":"32e7c7808153f8d8d5081ed3fa00ea4947824cd47e19028c84bf866954adf9ef"} Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.812564 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d4c82eb1-f999-45b7-82af-58e3b56a2e0a","Type":"ContainerStarted","Data":"528aad845f2a7b47b2a5bc924e7bdfb69195902a283f2d14900cfb11c4ec2a0b"} Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.812593 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d4c82eb1-f999-45b7-82af-58e3b56a2e0a","Type":"ContainerStarted","Data":"62722dd553f0f93dc04a5f464d231d41522dd12e4c2765f260f0264f09bee538"} Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.815321 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57d76178-7b70-4fc6-a865-579eed1fefe6","Type":"ContainerStarted","Data":"48faa8407d1a80b8cb5d75162a7b4cc7c709ba42bab0b764f583c54296bfd028"} Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.818286 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7be44442-eb08-4028-87d9-96decf3215b9","Type":"ContainerStarted","Data":"fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a"} Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.818327 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7be44442-eb08-4028-87d9-96decf3215b9","Type":"ContainerStarted","Data":"dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9"} Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.818359 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7be44442-eb08-4028-87d9-96decf3215b9","Type":"ContainerStarted","Data":"d432fbe5e4497cf2ad6644b72d2797da35d166bf3cc76328a0e6323d0de4d2cf"} Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.860706 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.860686562 podStartE2EDuration="2.860686562s" podCreationTimestamp="2025-12-11 09:00:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:26.860182946 +0000 UTC m=+1414.925598554" watchObservedRunningTime="2025-12-11 09:00:26.860686562 +0000 UTC m=+1414.926102180" Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.864936 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.864915715 podStartE2EDuration="2.864915715s" podCreationTimestamp="2025-12-11 09:00:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:26.838974181 +0000 UTC m=+1414.904389809" watchObservedRunningTime="2025-12-11 09:00:26.864915715 +0000 UTC m=+1414.930331333" Dec 11 09:00:26 crc kubenswrapper[4629]: I1211 09:00:26.879903 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.879883855 podStartE2EDuration="2.879883855s" podCreationTimestamp="2025-12-11 09:00:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:26.878068818 +0000 UTC m=+1414.943484436" watchObservedRunningTime="2025-12-11 09:00:26.879883855 +0000 UTC m=+1414.945299453" Dec 11 09:00:27 crc kubenswrapper[4629]: I1211 09:00:27.065892 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 11 09:00:28 crc kubenswrapper[4629]: I1211 09:00:28.855935 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57d76178-7b70-4fc6-a865-579eed1fefe6","Type":"ContainerStarted","Data":"b9f6b075e64d80065084263f75ba6763f68b59ba24387fd1e87df82448a4fbf1"} Dec 11 09:00:30 crc kubenswrapper[4629]: I1211 09:00:30.270378 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 11 09:00:30 crc kubenswrapper[4629]: I1211 09:00:30.384944 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 09:00:30 crc kubenswrapper[4629]: I1211 09:00:30.385060 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 09:00:31 crc kubenswrapper[4629]: I1211 09:00:31.911833 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57d76178-7b70-4fc6-a865-579eed1fefe6","Type":"ContainerStarted","Data":"5101adf5f34a1ed191bd9c2948fcc5768f83b4bde3a05834b344e79b5903e075"} Dec 11 09:00:32 crc kubenswrapper[4629]: I1211 09:00:32.044376 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 11 09:00:34 crc kubenswrapper[4629]: I1211 09:00:34.939212 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57d76178-7b70-4fc6-a865-579eed1fefe6","Type":"ContainerStarted","Data":"e55c9e1d51869475f00f64b195637ef1cb4fe424763f3ba3fb956292388a2bfa"} Dec 11 09:00:35 crc kubenswrapper[4629]: I1211 09:00:35.223561 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 09:00:35 crc kubenswrapper[4629]: I1211 09:00:35.223998 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 09:00:35 crc kubenswrapper[4629]: I1211 09:00:35.270010 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 11 09:00:35 crc kubenswrapper[4629]: I1211 09:00:35.295694 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 11 09:00:35 crc kubenswrapper[4629]: I1211 09:00:35.384555 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 11 09:00:35 crc kubenswrapper[4629]: I1211 09:00:35.384626 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 11 09:00:35 crc kubenswrapper[4629]: I1211 09:00:35.983999 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.528199295 podStartE2EDuration="11.983976357s" podCreationTimestamp="2025-12-11 09:00:24 +0000 UTC" firstStartedPulling="2025-12-11 09:00:25.052461412 +0000 UTC m=+1413.117877020" lastFinishedPulling="2025-12-11 09:00:34.508238454 +0000 UTC m=+1422.573654082" observedRunningTime="2025-12-11 09:00:35.97958446 +0000 UTC m=+1424.045000078" watchObservedRunningTime="2025-12-11 09:00:35.983976357 +0000 UTC m=+1424.049391965" Dec 11 09:00:35 crc kubenswrapper[4629]: I1211 09:00:35.994080 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 11 09:00:36 crc kubenswrapper[4629]: I1211 09:00:36.306439 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.176:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 09:00:36 crc kubenswrapper[4629]: I1211 09:00:36.306385 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.176:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 09:00:36 crc kubenswrapper[4629]: I1211 09:00:36.399120 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7be44442-eb08-4028-87d9-96decf3215b9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.178:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 09:00:36 crc kubenswrapper[4629]: I1211 09:00:36.399254 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7be44442-eb08-4028-87d9-96decf3215b9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.178:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 09:00:40 crc kubenswrapper[4629]: I1211 09:00:40.831087 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:40 crc kubenswrapper[4629]: I1211 09:00:40.954610 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-combined-ca-bundle\") pod \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\" (UID: \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\") " Dec 11 09:00:40 crc kubenswrapper[4629]: I1211 09:00:40.954701 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl29k\" (UniqueName: \"kubernetes.io/projected/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-kube-api-access-bl29k\") pod \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\" (UID: \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\") " Dec 11 09:00:40 crc kubenswrapper[4629]: I1211 09:00:40.954900 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-config-data\") pod \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\" (UID: \"b3204c2e-39a2-4fa3-a42e-33aa54521f0e\") " Dec 11 09:00:40 crc kubenswrapper[4629]: I1211 09:00:40.960766 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-kube-api-access-bl29k" (OuterVolumeSpecName: "kube-api-access-bl29k") pod "b3204c2e-39a2-4fa3-a42e-33aa54521f0e" (UID: "b3204c2e-39a2-4fa3-a42e-33aa54521f0e"). InnerVolumeSpecName "kube-api-access-bl29k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:40 crc kubenswrapper[4629]: I1211 09:00:40.986525 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-config-data" (OuterVolumeSpecName: "config-data") pod "b3204c2e-39a2-4fa3-a42e-33aa54521f0e" (UID: "b3204c2e-39a2-4fa3-a42e-33aa54521f0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:40 crc kubenswrapper[4629]: I1211 09:00:40.992584 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3204c2e-39a2-4fa3-a42e-33aa54521f0e" (UID: "b3204c2e-39a2-4fa3-a42e-33aa54521f0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:40 crc kubenswrapper[4629]: I1211 09:00:40.993650 4629 generic.go:334] "Generic (PLEG): container finished" podID="b3204c2e-39a2-4fa3-a42e-33aa54521f0e" containerID="d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3" exitCode=137 Dec 11 09:00:40 crc kubenswrapper[4629]: I1211 09:00:40.993690 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b3204c2e-39a2-4fa3-a42e-33aa54521f0e","Type":"ContainerDied","Data":"d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3"} Dec 11 09:00:40 crc kubenswrapper[4629]: I1211 09:00:40.993716 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b3204c2e-39a2-4fa3-a42e-33aa54521f0e","Type":"ContainerDied","Data":"68ba5b13116bd9e5ac20002414b0ea46735c3eace4f9c99621a285dbccd06abf"} Dec 11 09:00:40 crc kubenswrapper[4629]: I1211 09:00:40.993732 4629 scope.go:117] "RemoveContainer" containerID="d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3" Dec 11 09:00:40 crc kubenswrapper[4629]: I1211 09:00:40.993859 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.056980 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.057020 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.057070 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl29k\" (UniqueName: \"kubernetes.io/projected/b3204c2e-39a2-4fa3-a42e-33aa54521f0e-kube-api-access-bl29k\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.083856 4629 scope.go:117] "RemoveContainer" containerID="d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3" Dec 11 09:00:41 crc kubenswrapper[4629]: E1211 09:00:41.086432 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3\": container with ID starting with d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3 not found: ID does not exist" containerID="d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.086481 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3"} err="failed to get container status \"d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3\": rpc error: code = NotFound desc = could not find container \"d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3\": container with ID starting with d98617cca15a415f639a62480ef9184378fc1043242bff2c38073cc4998ab5f3 not found: ID does not exist" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.094943 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.105381 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.114171 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 09:00:41 crc kubenswrapper[4629]: E1211 09:00:41.114733 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3204c2e-39a2-4fa3-a42e-33aa54521f0e" containerName="nova-cell1-novncproxy-novncproxy" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.114756 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3204c2e-39a2-4fa3-a42e-33aa54521f0e" containerName="nova-cell1-novncproxy-novncproxy" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.115033 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3204c2e-39a2-4fa3-a42e-33aa54521f0e" containerName="nova-cell1-novncproxy-novncproxy" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.115977 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.119964 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.120315 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.120533 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.125531 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.260417 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.260502 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.260591 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwxpl\" (UniqueName: \"kubernetes.io/projected/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-kube-api-access-gwxpl\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.260618 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.260663 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.362868 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.363048 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.363086 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.363233 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwxpl\" (UniqueName: \"kubernetes.io/projected/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-kube-api-access-gwxpl\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.363322 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.367101 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.367160 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.367317 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.368457 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.382205 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwxpl\" (UniqueName: \"kubernetes.io/projected/82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7-kube-api-access-gwxpl\") pod \"nova-cell1-novncproxy-0\" (UID: \"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.454513 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:41 crc kubenswrapper[4629]: I1211 09:00:41.891108 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 09:00:42 crc kubenswrapper[4629]: I1211 09:00:42.026400 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7","Type":"ContainerStarted","Data":"257fef1c483924b491af651b930404fd790688ac3f1f184191f071dbd27dbbfd"} Dec 11 09:00:42 crc kubenswrapper[4629]: I1211 09:00:42.211260 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3204c2e-39a2-4fa3-a42e-33aa54521f0e" path="/var/lib/kubelet/pods/b3204c2e-39a2-4fa3-a42e-33aa54521f0e/volumes" Dec 11 09:00:43 crc kubenswrapper[4629]: I1211 09:00:43.037333 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7","Type":"ContainerStarted","Data":"21f0308b2c9954b0920720e90c915f38280c610991e0c21ad8abbc4ffd8a314e"} Dec 11 09:00:43 crc kubenswrapper[4629]: I1211 09:00:43.062784 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.062765518 podStartE2EDuration="2.062765518s" podCreationTimestamp="2025-12-11 09:00:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:43.06123721 +0000 UTC m=+1431.126652818" watchObservedRunningTime="2025-12-11 09:00:43.062765518 +0000 UTC m=+1431.128181126" Dec 11 09:00:45 crc kubenswrapper[4629]: I1211 09:00:45.230167 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 11 09:00:45 crc kubenswrapper[4629]: I1211 09:00:45.230639 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 11 09:00:45 crc kubenswrapper[4629]: I1211 09:00:45.235897 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 11 09:00:45 crc kubenswrapper[4629]: I1211 09:00:45.260905 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 11 09:00:45 crc kubenswrapper[4629]: I1211 09:00:45.390394 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 11 09:00:45 crc kubenswrapper[4629]: I1211 09:00:45.390745 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 11 09:00:45 crc kubenswrapper[4629]: I1211 09:00:45.398043 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 11 09:00:45 crc kubenswrapper[4629]: I1211 09:00:45.399162 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.060910 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.064991 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.283660 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-zkqrp"] Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.286166 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.325990 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-zkqrp"] Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.342857 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4xpfw"] Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.344784 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.411262 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4xpfw"] Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.455155 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.471922 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.472182 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.472321 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.472467 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-config\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.472543 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed4090d-b080-4da4-bb8d-bffd8a1680b2-catalog-content\") pod \"redhat-operators-4xpfw\" (UID: \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\") " pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.472633 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrztm\" (UniqueName: \"kubernetes.io/projected/d17744e4-f52a-4c2a-8a59-628d6e506f49-kube-api-access-jrztm\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.472730 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed4090d-b080-4da4-bb8d-bffd8a1680b2-utilities\") pod \"redhat-operators-4xpfw\" (UID: \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\") " pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.472820 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qj57\" (UniqueName: \"kubernetes.io/projected/aed4090d-b080-4da4-bb8d-bffd8a1680b2-kube-api-access-8qj57\") pod \"redhat-operators-4xpfw\" (UID: \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\") " pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.574149 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.574226 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.574361 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.574461 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-config\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.574488 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed4090d-b080-4da4-bb8d-bffd8a1680b2-catalog-content\") pod \"redhat-operators-4xpfw\" (UID: \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\") " pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.574513 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrztm\" (UniqueName: \"kubernetes.io/projected/d17744e4-f52a-4c2a-8a59-628d6e506f49-kube-api-access-jrztm\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.574554 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed4090d-b080-4da4-bb8d-bffd8a1680b2-utilities\") pod \"redhat-operators-4xpfw\" (UID: \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\") " pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.574606 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qj57\" (UniqueName: \"kubernetes.io/projected/aed4090d-b080-4da4-bb8d-bffd8a1680b2-kube-api-access-8qj57\") pod \"redhat-operators-4xpfw\" (UID: \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\") " pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.575216 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed4090d-b080-4da4-bb8d-bffd8a1680b2-utilities\") pod \"redhat-operators-4xpfw\" (UID: \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\") " pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.575274 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.575549 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed4090d-b080-4da4-bb8d-bffd8a1680b2-catalog-content\") pod \"redhat-operators-4xpfw\" (UID: \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\") " pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.575804 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-config\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.575949 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.576297 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.599915 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrztm\" (UniqueName: \"kubernetes.io/projected/d17744e4-f52a-4c2a-8a59-628d6e506f49-kube-api-access-jrztm\") pod \"dnsmasq-dns-68d4b6d797-zkqrp\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.602667 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qj57\" (UniqueName: \"kubernetes.io/projected/aed4090d-b080-4da4-bb8d-bffd8a1680b2-kube-api-access-8qj57\") pod \"redhat-operators-4xpfw\" (UID: \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\") " pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.648346 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:46 crc kubenswrapper[4629]: I1211 09:00:46.685382 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:00:47 crc kubenswrapper[4629]: I1211 09:00:47.237986 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-zkqrp"] Dec 11 09:00:47 crc kubenswrapper[4629]: I1211 09:00:47.399097 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4xpfw"] Dec 11 09:00:48 crc kubenswrapper[4629]: I1211 09:00:48.096839 4629 generic.go:334] "Generic (PLEG): container finished" podID="d17744e4-f52a-4c2a-8a59-628d6e506f49" containerID="ddd9af99d2ea9a8c5027dd523d564f83dc8b382932cdebccae7a4d8d8aa9933c" exitCode=0 Dec 11 09:00:48 crc kubenswrapper[4629]: I1211 09:00:48.097512 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" event={"ID":"d17744e4-f52a-4c2a-8a59-628d6e506f49","Type":"ContainerDied","Data":"ddd9af99d2ea9a8c5027dd523d564f83dc8b382932cdebccae7a4d8d8aa9933c"} Dec 11 09:00:48 crc kubenswrapper[4629]: I1211 09:00:48.097550 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" event={"ID":"d17744e4-f52a-4c2a-8a59-628d6e506f49","Type":"ContainerStarted","Data":"ddac5ffded1d4e1db29805322af71a6e528cd7264d9bf447fc122f4226ce67e4"} Dec 11 09:00:48 crc kubenswrapper[4629]: I1211 09:00:48.127603 4629 generic.go:334] "Generic (PLEG): container finished" podID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" containerID="bc907957c33da7c62363adf0334d5290dc1e2781513678ef7a444c18a0e12434" exitCode=0 Dec 11 09:00:48 crc kubenswrapper[4629]: I1211 09:00:48.128894 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xpfw" event={"ID":"aed4090d-b080-4da4-bb8d-bffd8a1680b2","Type":"ContainerDied","Data":"bc907957c33da7c62363adf0334d5290dc1e2781513678ef7a444c18a0e12434"} Dec 11 09:00:48 crc kubenswrapper[4629]: I1211 09:00:48.128928 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xpfw" event={"ID":"aed4090d-b080-4da4-bb8d-bffd8a1680b2","Type":"ContainerStarted","Data":"e20cf1a3746663723eed2738908e1dc7491d20a6768b44caff04b5e8d3dff9c2"} Dec 11 09:00:49 crc kubenswrapper[4629]: I1211 09:00:49.143417 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" event={"ID":"d17744e4-f52a-4c2a-8a59-628d6e506f49","Type":"ContainerStarted","Data":"c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915"} Dec 11 09:00:49 crc kubenswrapper[4629]: I1211 09:00:49.143719 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:49 crc kubenswrapper[4629]: I1211 09:00:49.180040 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" podStartSLOduration=3.180023375 podStartE2EDuration="3.180023375s" podCreationTimestamp="2025-12-11 09:00:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:49.171931731 +0000 UTC m=+1437.237347339" watchObservedRunningTime="2025-12-11 09:00:49.180023375 +0000 UTC m=+1437.245438983" Dec 11 09:00:49 crc kubenswrapper[4629]: I1211 09:00:49.472450 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:49 crc kubenswrapper[4629]: I1211 09:00:49.472705 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" containerName="nova-api-log" containerID="cri-o://2d9b89d6eb5a1a43bb565dbbe1f5ce122c1bd5c131653d9eac3eeaf3c39f9d7a" gracePeriod=30 Dec 11 09:00:49 crc kubenswrapper[4629]: I1211 09:00:49.472794 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" containerName="nova-api-api" containerID="cri-o://83cd8eeec150772c31861c882ccceed6aa5c21377536575900682c7209bea2eb" gracePeriod=30 Dec 11 09:00:51 crc kubenswrapper[4629]: I1211 09:00:51.104957 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 09:00:51 crc kubenswrapper[4629]: I1211 09:00:51.107148 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 09:00:51 crc kubenswrapper[4629]: I1211 09:00:51.107208 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="proxy-httpd" containerID="cri-o://e55c9e1d51869475f00f64b195637ef1cb4fe424763f3ba3fb956292388a2bfa" gracePeriod=30 Dec 11 09:00:51 crc kubenswrapper[4629]: I1211 09:00:51.107351 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="sg-core" containerID="cri-o://5101adf5f34a1ed191bd9c2948fcc5768f83b4bde3a05834b344e79b5903e075" gracePeriod=30 Dec 11 09:00:51 crc kubenswrapper[4629]: I1211 09:00:51.107398 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="ceilometer-notification-agent" containerID="cri-o://b9f6b075e64d80065084263f75ba6763f68b59ba24387fd1e87df82448a4fbf1" gracePeriod=30 Dec 11 09:00:51 crc kubenswrapper[4629]: I1211 09:00:51.108887 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="ceilometer-central-agent" containerID="cri-o://48faa8407d1a80b8cb5d75162a7b4cc7c709ba42bab0b764f583c54296bfd028" gracePeriod=30 Dec 11 09:00:51 crc kubenswrapper[4629]: I1211 09:00:51.143448 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.175:3000/\": EOF" Dec 11 09:00:51 crc kubenswrapper[4629]: I1211 09:00:51.186627 4629 generic.go:334] "Generic (PLEG): container finished" podID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" containerID="2d9b89d6eb5a1a43bb565dbbe1f5ce122c1bd5c131653d9eac3eeaf3c39f9d7a" exitCode=143 Dec 11 09:00:51 crc kubenswrapper[4629]: I1211 09:00:51.186690 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105","Type":"ContainerDied","Data":"2d9b89d6eb5a1a43bb565dbbe1f5ce122c1bd5c131653d9eac3eeaf3c39f9d7a"} Dec 11 09:00:51 crc kubenswrapper[4629]: I1211 09:00:51.455108 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:51 crc kubenswrapper[4629]: I1211 09:00:51.475781 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.234152 4629 generic.go:334] "Generic (PLEG): container finished" podID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerID="e55c9e1d51869475f00f64b195637ef1cb4fe424763f3ba3fb956292388a2bfa" exitCode=0 Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.234195 4629 generic.go:334] "Generic (PLEG): container finished" podID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerID="5101adf5f34a1ed191bd9c2948fcc5768f83b4bde3a05834b344e79b5903e075" exitCode=2 Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.234210 4629 generic.go:334] "Generic (PLEG): container finished" podID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerID="48faa8407d1a80b8cb5d75162a7b4cc7c709ba42bab0b764f583c54296bfd028" exitCode=0 Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.234344 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57d76178-7b70-4fc6-a865-579eed1fefe6","Type":"ContainerDied","Data":"e55c9e1d51869475f00f64b195637ef1cb4fe424763f3ba3fb956292388a2bfa"} Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.234382 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57d76178-7b70-4fc6-a865-579eed1fefe6","Type":"ContainerDied","Data":"5101adf5f34a1ed191bd9c2948fcc5768f83b4bde3a05834b344e79b5903e075"} Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.234398 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57d76178-7b70-4fc6-a865-579eed1fefe6","Type":"ContainerDied","Data":"48faa8407d1a80b8cb5d75162a7b4cc7c709ba42bab0b764f583c54296bfd028"} Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.246929 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xpfw" event={"ID":"aed4090d-b080-4da4-bb8d-bffd8a1680b2","Type":"ContainerStarted","Data":"87df4abdc5f28de3fd89dba33399049ee90091f7be28376c7edec2b6879edf14"} Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.385262 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.813198 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-nmjwk"] Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.816207 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.818885 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.819195 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.846167 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-nmjwk"] Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.960150 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g57rk\" (UniqueName: \"kubernetes.io/projected/2de77a96-4287-46cc-8004-04f26169a35d-kube-api-access-g57rk\") pod \"nova-cell1-cell-mapping-nmjwk\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.960230 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-scripts\") pod \"nova-cell1-cell-mapping-nmjwk\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.960663 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-config-data\") pod \"nova-cell1-cell-mapping-nmjwk\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:52 crc kubenswrapper[4629]: I1211 09:00:52.961025 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-nmjwk\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:53 crc kubenswrapper[4629]: I1211 09:00:53.062506 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-scripts\") pod \"nova-cell1-cell-mapping-nmjwk\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:53 crc kubenswrapper[4629]: I1211 09:00:53.062609 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-config-data\") pod \"nova-cell1-cell-mapping-nmjwk\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:53 crc kubenswrapper[4629]: I1211 09:00:53.062689 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-nmjwk\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:53 crc kubenswrapper[4629]: I1211 09:00:53.062750 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g57rk\" (UniqueName: \"kubernetes.io/projected/2de77a96-4287-46cc-8004-04f26169a35d-kube-api-access-g57rk\") pod \"nova-cell1-cell-mapping-nmjwk\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:53 crc kubenswrapper[4629]: I1211 09:00:53.070505 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-scripts\") pod \"nova-cell1-cell-mapping-nmjwk\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:53 crc kubenswrapper[4629]: I1211 09:00:53.070758 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-config-data\") pod \"nova-cell1-cell-mapping-nmjwk\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:53 crc kubenswrapper[4629]: I1211 09:00:53.077019 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-nmjwk\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:53 crc kubenswrapper[4629]: I1211 09:00:53.082800 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g57rk\" (UniqueName: \"kubernetes.io/projected/2de77a96-4287-46cc-8004-04f26169a35d-kube-api-access-g57rk\") pod \"nova-cell1-cell-mapping-nmjwk\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:53 crc kubenswrapper[4629]: I1211 09:00:53.161504 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:00:53 crc kubenswrapper[4629]: W1211 09:00:53.845804 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2de77a96_4287_46cc_8004_04f26169a35d.slice/crio-0a604d6a00b0a79fb98b3f28d11eb548672a006941111c7f0f3f8bf090e53584 WatchSource:0}: Error finding container 0a604d6a00b0a79fb98b3f28d11eb548672a006941111c7f0f3f8bf090e53584: Status 404 returned error can't find the container with id 0a604d6a00b0a79fb98b3f28d11eb548672a006941111c7f0f3f8bf090e53584 Dec 11 09:00:53 crc kubenswrapper[4629]: I1211 09:00:53.855253 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-nmjwk"] Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.286616 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-nmjwk" event={"ID":"2de77a96-4287-46cc-8004-04f26169a35d","Type":"ContainerStarted","Data":"7167f4bf289d9b5cd36d9756beb18bfb61651fce14cb268b6be59134ddb91ded"} Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.287017 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-nmjwk" event={"ID":"2de77a96-4287-46cc-8004-04f26169a35d","Type":"ContainerStarted","Data":"0a604d6a00b0a79fb98b3f28d11eb548672a006941111c7f0f3f8bf090e53584"} Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.294654 4629 generic.go:334] "Generic (PLEG): container finished" podID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" containerID="83cd8eeec150772c31861c882ccceed6aa5c21377536575900682c7209bea2eb" exitCode=0 Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.294707 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105","Type":"ContainerDied","Data":"83cd8eeec150772c31861c882ccceed6aa5c21377536575900682c7209bea2eb"} Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.322829 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-nmjwk" podStartSLOduration=2.322809991 podStartE2EDuration="2.322809991s" podCreationTimestamp="2025-12-11 09:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:54.306104327 +0000 UTC m=+1442.371519935" watchObservedRunningTime="2025-12-11 09:00:54.322809991 +0000 UTC m=+1442.388225599" Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.329464 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.393320 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-combined-ca-bundle\") pod \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.393616 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-config-data\") pod \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.394207 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-logs\") pod \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.394368 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzmxn\" (UniqueName: \"kubernetes.io/projected/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-kube-api-access-fzmxn\") pod \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\" (UID: \"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105\") " Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.395763 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-logs" (OuterVolumeSpecName: "logs") pod "cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" (UID: "cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.399655 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-kube-api-access-fzmxn" (OuterVolumeSpecName: "kube-api-access-fzmxn") pod "cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" (UID: "cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105"). InnerVolumeSpecName "kube-api-access-fzmxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.435965 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.175:3000/\": dial tcp 10.217.0.175:3000: connect: connection refused" Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.438498 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" (UID: "cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.451026 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-config-data" (OuterVolumeSpecName: "config-data") pod "cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" (UID: "cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.498203 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.498236 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.498247 4629 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-logs\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:54 crc kubenswrapper[4629]: I1211 09:00:54.498256 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzmxn\" (UniqueName: \"kubernetes.io/projected/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105-kube-api-access-fzmxn\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.309950 4629 generic.go:334] "Generic (PLEG): container finished" podID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" containerID="87df4abdc5f28de3fd89dba33399049ee90091f7be28376c7edec2b6879edf14" exitCode=0 Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.310325 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xpfw" event={"ID":"aed4090d-b080-4da4-bb8d-bffd8a1680b2","Type":"ContainerDied","Data":"87df4abdc5f28de3fd89dba33399049ee90091f7be28376c7edec2b6879edf14"} Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.315764 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105","Type":"ContainerDied","Data":"32e7c7808153f8d8d5081ed3fa00ea4947824cd47e19028c84bf866954adf9ef"} Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.315791 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.315827 4629 scope.go:117] "RemoveContainer" containerID="83cd8eeec150772c31861c882ccceed6aa5c21377536575900682c7209bea2eb" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.353972 4629 scope.go:117] "RemoveContainer" containerID="2d9b89d6eb5a1a43bb565dbbe1f5ce122c1bd5c131653d9eac3eeaf3c39f9d7a" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.396915 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.409244 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.419313 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:55 crc kubenswrapper[4629]: E1211 09:00:55.419752 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" containerName="nova-api-log" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.419774 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" containerName="nova-api-log" Dec 11 09:00:55 crc kubenswrapper[4629]: E1211 09:00:55.419818 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" containerName="nova-api-api" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.419825 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" containerName="nova-api-api" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.420057 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" containerName="nova-api-log" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.420093 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" containerName="nova-api-api" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.421240 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.425273 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.428871 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.429040 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.516395 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s86p\" (UniqueName: \"kubernetes.io/projected/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-kube-api-access-9s86p\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.516698 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-config-data\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.516824 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-public-tls-certs\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.516979 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.517107 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.517272 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-logs\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.618690 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.619146 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.619304 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-logs\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.619531 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s86p\" (UniqueName: \"kubernetes.io/projected/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-kube-api-access-9s86p\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.619758 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-config-data\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.619903 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-public-tls-certs\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.620117 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-logs\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.624352 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.624387 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.628554 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-config-data\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.631211 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-public-tls-certs\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.637351 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s86p\" (UniqueName: \"kubernetes.io/projected/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-kube-api-access-9s86p\") pod \"nova-api-0\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " pod="openstack/nova-api-0" Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.687036 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:55 crc kubenswrapper[4629]: I1211 09:00:55.756746 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.211498 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105" path="/var/lib/kubelet/pods/cd7cca7d-28db-4a9e-aca8-bd8e0e6fd105/volumes" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.267110 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.334137 4629 generic.go:334] "Generic (PLEG): container finished" podID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerID="b9f6b075e64d80065084263f75ba6763f68b59ba24387fd1e87df82448a4fbf1" exitCode=0 Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.334218 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57d76178-7b70-4fc6-a865-579eed1fefe6","Type":"ContainerDied","Data":"b9f6b075e64d80065084263f75ba6763f68b59ba24387fd1e87df82448a4fbf1"} Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.337055 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1c37cee-a766-4bae-bcc7-a6a69f2876f1","Type":"ContainerStarted","Data":"60e04aa43945fd26c655a186a90c722a866eba54c3700ed0c6ffae03fa38ca8a"} Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.528416 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.649952 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.653683 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lss5l\" (UniqueName: \"kubernetes.io/projected/57d76178-7b70-4fc6-a865-579eed1fefe6-kube-api-access-lss5l\") pod \"57d76178-7b70-4fc6-a865-579eed1fefe6\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.653745 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-scripts\") pod \"57d76178-7b70-4fc6-a865-579eed1fefe6\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.653821 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-combined-ca-bundle\") pod \"57d76178-7b70-4fc6-a865-579eed1fefe6\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.653921 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-config-data\") pod \"57d76178-7b70-4fc6-a865-579eed1fefe6\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.653975 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57d76178-7b70-4fc6-a865-579eed1fefe6-log-httpd\") pod \"57d76178-7b70-4fc6-a865-579eed1fefe6\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.654033 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-sg-core-conf-yaml\") pod \"57d76178-7b70-4fc6-a865-579eed1fefe6\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.654091 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-ceilometer-tls-certs\") pod \"57d76178-7b70-4fc6-a865-579eed1fefe6\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.654132 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57d76178-7b70-4fc6-a865-579eed1fefe6-run-httpd\") pod \"57d76178-7b70-4fc6-a865-579eed1fefe6\" (UID: \"57d76178-7b70-4fc6-a865-579eed1fefe6\") " Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.655097 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57d76178-7b70-4fc6-a865-579eed1fefe6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "57d76178-7b70-4fc6-a865-579eed1fefe6" (UID: "57d76178-7b70-4fc6-a865-579eed1fefe6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.655491 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57d76178-7b70-4fc6-a865-579eed1fefe6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "57d76178-7b70-4fc6-a865-579eed1fefe6" (UID: "57d76178-7b70-4fc6-a865-579eed1fefe6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.663328 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57d76178-7b70-4fc6-a865-579eed1fefe6-kube-api-access-lss5l" (OuterVolumeSpecName: "kube-api-access-lss5l") pod "57d76178-7b70-4fc6-a865-579eed1fefe6" (UID: "57d76178-7b70-4fc6-a865-579eed1fefe6"). InnerVolumeSpecName "kube-api-access-lss5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.705101 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-scripts" (OuterVolumeSpecName: "scripts") pod "57d76178-7b70-4fc6-a865-579eed1fefe6" (UID: "57d76178-7b70-4fc6-a865-579eed1fefe6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.734745 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-55m6h"] Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.735044 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" podUID="5c29fe70-580c-4f60-9dbc-d9bae34e376d" containerName="dnsmasq-dns" containerID="cri-o://8cdd2e499d8a8d46810c4f6a3b338bc25758cb8545afb5aca32df8a634eabb75" gracePeriod=10 Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.745074 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "57d76178-7b70-4fc6-a865-579eed1fefe6" (UID: "57d76178-7b70-4fc6-a865-579eed1fefe6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.773082 4629 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57d76178-7b70-4fc6-a865-579eed1fefe6-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.773125 4629 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.773140 4629 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57d76178-7b70-4fc6-a865-579eed1fefe6-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.773152 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lss5l\" (UniqueName: \"kubernetes.io/projected/57d76178-7b70-4fc6-a865-579eed1fefe6-kube-api-access-lss5l\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.773162 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.848333 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57d76178-7b70-4fc6-a865-579eed1fefe6" (UID: "57d76178-7b70-4fc6-a865-579eed1fefe6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.874899 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "57d76178-7b70-4fc6-a865-579eed1fefe6" (UID: "57d76178-7b70-4fc6-a865-579eed1fefe6"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.876275 4629 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.876302 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.923501 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-config-data" (OuterVolumeSpecName: "config-data") pod "57d76178-7b70-4fc6-a865-579eed1fefe6" (UID: "57d76178-7b70-4fc6-a865-579eed1fefe6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:56 crc kubenswrapper[4629]: I1211 09:00:56.978268 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57d76178-7b70-4fc6-a865-579eed1fefe6-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.351562 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.351562 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57d76178-7b70-4fc6-a865-579eed1fefe6","Type":"ContainerDied","Data":"0fcd6c252e8de703ba3100cd6c40f0704e2e26161f2ade5ef6cede8e6e736a64"} Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.352023 4629 scope.go:117] "RemoveContainer" containerID="e55c9e1d51869475f00f64b195637ef1cb4fe424763f3ba3fb956292388a2bfa" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.353499 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1c37cee-a766-4bae-bcc7-a6a69f2876f1","Type":"ContainerStarted","Data":"5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd"} Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.353542 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1c37cee-a766-4bae-bcc7-a6a69f2876f1","Type":"ContainerStarted","Data":"de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141"} Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.371083 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xpfw" event={"ID":"aed4090d-b080-4da4-bb8d-bffd8a1680b2","Type":"ContainerStarted","Data":"650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f"} Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.377505 4629 scope.go:117] "RemoveContainer" containerID="5101adf5f34a1ed191bd9c2948fcc5768f83b4bde3a05834b344e79b5903e075" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.382887 4629 generic.go:334] "Generic (PLEG): container finished" podID="5c29fe70-580c-4f60-9dbc-d9bae34e376d" containerID="8cdd2e499d8a8d46810c4f6a3b338bc25758cb8545afb5aca32df8a634eabb75" exitCode=0 Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.382932 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" event={"ID":"5c29fe70-580c-4f60-9dbc-d9bae34e376d","Type":"ContainerDied","Data":"8cdd2e499d8a8d46810c4f6a3b338bc25758cb8545afb5aca32df8a634eabb75"} Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.382957 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" event={"ID":"5c29fe70-580c-4f60-9dbc-d9bae34e376d","Type":"ContainerDied","Data":"4accc909aee8920af1e2c3ece6d4d8ae07646abbdf0bbbe7fb8fc16118ca9bc6"} Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.382968 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4accc909aee8920af1e2c3ece6d4d8ae07646abbdf0bbbe7fb8fc16118ca9bc6" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.384063 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.410722 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.410701439 podStartE2EDuration="2.410701439s" podCreationTimestamp="2025-12-11 09:00:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:00:57.390266617 +0000 UTC m=+1445.455682215" watchObservedRunningTime="2025-12-11 09:00:57.410701439 +0000 UTC m=+1445.476117047" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.429745 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4xpfw" podStartSLOduration=2.9068988989999998 podStartE2EDuration="11.429725916s" podCreationTimestamp="2025-12-11 09:00:46 +0000 UTC" firstStartedPulling="2025-12-11 09:00:48.130081505 +0000 UTC m=+1436.195497113" lastFinishedPulling="2025-12-11 09:00:56.652908532 +0000 UTC m=+1444.718324130" observedRunningTime="2025-12-11 09:00:57.420292619 +0000 UTC m=+1445.485708227" watchObservedRunningTime="2025-12-11 09:00:57.429725916 +0000 UTC m=+1445.495141524" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.433155 4629 scope.go:117] "RemoveContainer" containerID="b9f6b075e64d80065084263f75ba6763f68b59ba24387fd1e87df82448a4fbf1" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.457142 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.469095 4629 scope.go:117] "RemoveContainer" containerID="48faa8407d1a80b8cb5d75162a7b4cc7c709ba42bab0b764f583c54296bfd028" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.473580 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.496036 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q5n2\" (UniqueName: \"kubernetes.io/projected/5c29fe70-580c-4f60-9dbc-d9bae34e376d-kube-api-access-7q5n2\") pod \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.496289 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-ovsdbserver-sb\") pod \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.496348 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-ovsdbserver-nb\") pod \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.496453 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-config\") pod \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.496483 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-dns-svc\") pod \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\" (UID: \"5c29fe70-580c-4f60-9dbc-d9bae34e376d\") " Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.547127 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 09:00:57 crc kubenswrapper[4629]: E1211 09:00:57.548908 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="ceilometer-notification-agent" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.548931 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="ceilometer-notification-agent" Dec 11 09:00:57 crc kubenswrapper[4629]: E1211 09:00:57.548962 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c29fe70-580c-4f60-9dbc-d9bae34e376d" containerName="init" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.548968 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c29fe70-580c-4f60-9dbc-d9bae34e376d" containerName="init" Dec 11 09:00:57 crc kubenswrapper[4629]: E1211 09:00:57.548986 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c29fe70-580c-4f60-9dbc-d9bae34e376d" containerName="dnsmasq-dns" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.548993 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c29fe70-580c-4f60-9dbc-d9bae34e376d" containerName="dnsmasq-dns" Dec 11 09:00:57 crc kubenswrapper[4629]: E1211 09:00:57.549026 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="ceilometer-central-agent" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.549035 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="ceilometer-central-agent" Dec 11 09:00:57 crc kubenswrapper[4629]: E1211 09:00:57.549045 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="sg-core" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.549051 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="sg-core" Dec 11 09:00:57 crc kubenswrapper[4629]: E1211 09:00:57.549075 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="proxy-httpd" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.549081 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="proxy-httpd" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.549523 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="ceilometer-notification-agent" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.549543 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c29fe70-580c-4f60-9dbc-d9bae34e376d" containerName="dnsmasq-dns" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.549567 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="proxy-httpd" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.549595 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="ceilometer-central-agent" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.549613 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" containerName="sg-core" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.564126 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c29fe70-580c-4f60-9dbc-d9bae34e376d-kube-api-access-7q5n2" (OuterVolumeSpecName: "kube-api-access-7q5n2") pod "5c29fe70-580c-4f60-9dbc-d9bae34e376d" (UID: "5c29fe70-580c-4f60-9dbc-d9bae34e376d"). InnerVolumeSpecName "kube-api-access-7q5n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.564614 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.574734 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.574995 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.575105 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.602762 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q5n2\" (UniqueName: \"kubernetes.io/projected/5c29fe70-580c-4f60-9dbc-d9bae34e376d-kube-api-access-7q5n2\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.613229 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.637128 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5c29fe70-580c-4f60-9dbc-d9bae34e376d" (UID: "5c29fe70-580c-4f60-9dbc-d9bae34e376d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.656300 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5c29fe70-580c-4f60-9dbc-d9bae34e376d" (UID: "5c29fe70-580c-4f60-9dbc-d9bae34e376d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.660214 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-config" (OuterVolumeSpecName: "config") pod "5c29fe70-580c-4f60-9dbc-d9bae34e376d" (UID: "5c29fe70-580c-4f60-9dbc-d9bae34e376d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.668468 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5c29fe70-580c-4f60-9dbc-d9bae34e376d" (UID: "5c29fe70-580c-4f60-9dbc-d9bae34e376d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.704373 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.704441 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-config-data\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.704468 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-scripts\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.704496 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9568824-b16c-4084-90c7-dd6fe60a866a-log-httpd\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.704667 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.704693 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9568824-b16c-4084-90c7-dd6fe60a866a-run-httpd\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.704728 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.704754 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqgjv\" (UniqueName: \"kubernetes.io/projected/a9568824-b16c-4084-90c7-dd6fe60a866a-kube-api-access-wqgjv\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.704871 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.704891 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.704903 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-config\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.704915 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c29fe70-580c-4f60-9dbc-d9bae34e376d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.806708 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.806782 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqgjv\" (UniqueName: \"kubernetes.io/projected/a9568824-b16c-4084-90c7-dd6fe60a866a-kube-api-access-wqgjv\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.806866 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.806886 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-config-data\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.806906 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-scripts\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.806923 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9568824-b16c-4084-90c7-dd6fe60a866a-log-httpd\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.807038 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.807054 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9568824-b16c-4084-90c7-dd6fe60a866a-run-httpd\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.807511 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9568824-b16c-4084-90c7-dd6fe60a866a-run-httpd\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.809132 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9568824-b16c-4084-90c7-dd6fe60a866a-log-httpd\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.814131 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.815754 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-scripts\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.816150 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.817521 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.818362 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9568824-b16c-4084-90c7-dd6fe60a866a-config-data\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.829358 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqgjv\" (UniqueName: \"kubernetes.io/projected/a9568824-b16c-4084-90c7-dd6fe60a866a-kube-api-access-wqgjv\") pod \"ceilometer-0\" (UID: \"a9568824-b16c-4084-90c7-dd6fe60a866a\") " pod="openstack/ceilometer-0" Dec 11 09:00:57 crc kubenswrapper[4629]: I1211 09:00:57.909465 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 09:00:58 crc kubenswrapper[4629]: I1211 09:00:58.211292 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57d76178-7b70-4fc6-a865-579eed1fefe6" path="/var/lib/kubelet/pods/57d76178-7b70-4fc6-a865-579eed1fefe6/volumes" Dec 11 09:00:58 crc kubenswrapper[4629]: I1211 09:00:58.394053 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-55m6h" Dec 11 09:00:58 crc kubenswrapper[4629]: I1211 09:00:58.418697 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-55m6h"] Dec 11 09:00:58 crc kubenswrapper[4629]: I1211 09:00:58.426997 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-55m6h"] Dec 11 09:00:58 crc kubenswrapper[4629]: I1211 09:00:58.438953 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 09:00:59 crc kubenswrapper[4629]: I1211 09:00:59.404255 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9568824-b16c-4084-90c7-dd6fe60a866a","Type":"ContainerStarted","Data":"86cdfc3e0cf8d98199f16739f0704f18861c1d3808fb84cea7b120d7bfe86484"} Dec 11 09:00:59 crc kubenswrapper[4629]: I1211 09:00:59.404626 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9568824-b16c-4084-90c7-dd6fe60a866a","Type":"ContainerStarted","Data":"784ffb210de48f3c0127580eb08576d0cb6a83e7c4bac86a07d734108c2e458f"} Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.148169 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29424061-mzwc9"] Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.149434 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.168035 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29424061-mzwc9"] Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.275359 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c29fe70-580c-4f60-9dbc-d9bae34e376d" path="/var/lib/kubelet/pods/5c29fe70-580c-4f60-9dbc-d9bae34e376d/volumes" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.363813 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rplg5\" (UniqueName: \"kubernetes.io/projected/cef5a8b9-b204-44a9-92fd-5bd1405895dd-kube-api-access-rplg5\") pod \"keystone-cron-29424061-mzwc9\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.363907 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-fernet-keys\") pod \"keystone-cron-29424061-mzwc9\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.363957 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-config-data\") pod \"keystone-cron-29424061-mzwc9\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.363975 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-combined-ca-bundle\") pod \"keystone-cron-29424061-mzwc9\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.465726 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rplg5\" (UniqueName: \"kubernetes.io/projected/cef5a8b9-b204-44a9-92fd-5bd1405895dd-kube-api-access-rplg5\") pod \"keystone-cron-29424061-mzwc9\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.465836 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-fernet-keys\") pod \"keystone-cron-29424061-mzwc9\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.465955 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-config-data\") pod \"keystone-cron-29424061-mzwc9\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.465982 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-combined-ca-bundle\") pod \"keystone-cron-29424061-mzwc9\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.470170 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-config-data\") pod \"keystone-cron-29424061-mzwc9\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.470786 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-fernet-keys\") pod \"keystone-cron-29424061-mzwc9\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.471645 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-combined-ca-bundle\") pod \"keystone-cron-29424061-mzwc9\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.489352 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rplg5\" (UniqueName: \"kubernetes.io/projected/cef5a8b9-b204-44a9-92fd-5bd1405895dd-kube-api-access-rplg5\") pod \"keystone-cron-29424061-mzwc9\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:00 crc kubenswrapper[4629]: I1211 09:01:00.580897 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:01 crc kubenswrapper[4629]: I1211 09:01:01.086955 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29424061-mzwc9"] Dec 11 09:01:01 crc kubenswrapper[4629]: I1211 09:01:01.429929 4629 generic.go:334] "Generic (PLEG): container finished" podID="2de77a96-4287-46cc-8004-04f26169a35d" containerID="7167f4bf289d9b5cd36d9756beb18bfb61651fce14cb268b6be59134ddb91ded" exitCode=0 Dec 11 09:01:01 crc kubenswrapper[4629]: I1211 09:01:01.430026 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-nmjwk" event={"ID":"2de77a96-4287-46cc-8004-04f26169a35d","Type":"ContainerDied","Data":"7167f4bf289d9b5cd36d9756beb18bfb61651fce14cb268b6be59134ddb91ded"} Dec 11 09:01:01 crc kubenswrapper[4629]: I1211 09:01:01.432781 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9568824-b16c-4084-90c7-dd6fe60a866a","Type":"ContainerStarted","Data":"a32ef7f517d78be267135290424ec6010e674ef87348d3a360b30f58f04facfd"} Dec 11 09:01:01 crc kubenswrapper[4629]: I1211 09:01:01.435407 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29424061-mzwc9" event={"ID":"cef5a8b9-b204-44a9-92fd-5bd1405895dd","Type":"ContainerStarted","Data":"fc5af8726f2434e79defccb71eecf9c7055a8afef5488f5f6077a554ef738875"} Dec 11 09:01:01 crc kubenswrapper[4629]: I1211 09:01:01.435444 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29424061-mzwc9" event={"ID":"cef5a8b9-b204-44a9-92fd-5bd1405895dd","Type":"ContainerStarted","Data":"dc9f7255ccae1694425834ed1bb9beabe18e586dee92ff0006f265d32c010c70"} Dec 11 09:01:01 crc kubenswrapper[4629]: I1211 09:01:01.474571 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29424061-mzwc9" podStartSLOduration=1.474553323 podStartE2EDuration="1.474553323s" podCreationTimestamp="2025-12-11 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:01:01.468281107 +0000 UTC m=+1449.533696725" watchObservedRunningTime="2025-12-11 09:01:01.474553323 +0000 UTC m=+1449.539968931" Dec 11 09:01:02 crc kubenswrapper[4629]: I1211 09:01:02.456674 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9568824-b16c-4084-90c7-dd6fe60a866a","Type":"ContainerStarted","Data":"5eff6545be93f71823c292c8e73bfc5d55ae85212fb571cc26bf84835d47505b"} Dec 11 09:01:02 crc kubenswrapper[4629]: I1211 09:01:02.830223 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.015550 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-scripts\") pod \"2de77a96-4287-46cc-8004-04f26169a35d\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.015660 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-combined-ca-bundle\") pod \"2de77a96-4287-46cc-8004-04f26169a35d\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.015754 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-config-data\") pod \"2de77a96-4287-46cc-8004-04f26169a35d\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.015891 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g57rk\" (UniqueName: \"kubernetes.io/projected/2de77a96-4287-46cc-8004-04f26169a35d-kube-api-access-g57rk\") pod \"2de77a96-4287-46cc-8004-04f26169a35d\" (UID: \"2de77a96-4287-46cc-8004-04f26169a35d\") " Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.033598 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-scripts" (OuterVolumeSpecName: "scripts") pod "2de77a96-4287-46cc-8004-04f26169a35d" (UID: "2de77a96-4287-46cc-8004-04f26169a35d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.038046 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2de77a96-4287-46cc-8004-04f26169a35d-kube-api-access-g57rk" (OuterVolumeSpecName: "kube-api-access-g57rk") pod "2de77a96-4287-46cc-8004-04f26169a35d" (UID: "2de77a96-4287-46cc-8004-04f26169a35d"). InnerVolumeSpecName "kube-api-access-g57rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.063967 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2de77a96-4287-46cc-8004-04f26169a35d" (UID: "2de77a96-4287-46cc-8004-04f26169a35d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.071128 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-config-data" (OuterVolumeSpecName: "config-data") pod "2de77a96-4287-46cc-8004-04f26169a35d" (UID: "2de77a96-4287-46cc-8004-04f26169a35d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.119237 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.119270 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g57rk\" (UniqueName: \"kubernetes.io/projected/2de77a96-4287-46cc-8004-04f26169a35d-kube-api-access-g57rk\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.119281 4629 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.119293 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de77a96-4287-46cc-8004-04f26169a35d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.468678 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-nmjwk" event={"ID":"2de77a96-4287-46cc-8004-04f26169a35d","Type":"ContainerDied","Data":"0a604d6a00b0a79fb98b3f28d11eb548672a006941111c7f0f3f8bf090e53584"} Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.469050 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a604d6a00b0a79fb98b3f28d11eb548672a006941111c7f0f3f8bf090e53584" Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.468757 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-nmjwk" Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.633949 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.634248 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c1c37cee-a766-4bae-bcc7-a6a69f2876f1" containerName="nova-api-log" containerID="cri-o://de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141" gracePeriod=30 Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.634322 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c1c37cee-a766-4bae-bcc7-a6a69f2876f1" containerName="nova-api-api" containerID="cri-o://5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd" gracePeriod=30 Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.652299 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.653500 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d4c82eb1-f999-45b7-82af-58e3b56a2e0a" containerName="nova-scheduler-scheduler" containerID="cri-o://528aad845f2a7b47b2a5bc924e7bdfb69195902a283f2d14900cfb11c4ec2a0b" gracePeriod=30 Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.675241 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.675473 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7be44442-eb08-4028-87d9-96decf3215b9" containerName="nova-metadata-log" containerID="cri-o://dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9" gracePeriod=30 Dec 11 09:01:03 crc kubenswrapper[4629]: I1211 09:01:03.675936 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7be44442-eb08-4028-87d9-96decf3215b9" containerName="nova-metadata-metadata" containerID="cri-o://fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a" gracePeriod=30 Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.333929 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.352120 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-combined-ca-bundle\") pod \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.352197 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-public-tls-certs\") pod \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.352287 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-logs\") pod \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.352744 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-internal-tls-certs\") pod \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.352675 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-logs" (OuterVolumeSpecName: "logs") pod "c1c37cee-a766-4bae-bcc7-a6a69f2876f1" (UID: "c1c37cee-a766-4bae-bcc7-a6a69f2876f1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.353172 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s86p\" (UniqueName: \"kubernetes.io/projected/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-kube-api-access-9s86p\") pod \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.353244 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-config-data\") pod \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\" (UID: \"c1c37cee-a766-4bae-bcc7-a6a69f2876f1\") " Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.353972 4629 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-logs\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.364658 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-kube-api-access-9s86p" (OuterVolumeSpecName: "kube-api-access-9s86p") pod "c1c37cee-a766-4bae-bcc7-a6a69f2876f1" (UID: "c1c37cee-a766-4bae-bcc7-a6a69f2876f1"). InnerVolumeSpecName "kube-api-access-9s86p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.392119 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1c37cee-a766-4bae-bcc7-a6a69f2876f1" (UID: "c1c37cee-a766-4bae-bcc7-a6a69f2876f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.425653 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-config-data" (OuterVolumeSpecName: "config-data") pod "c1c37cee-a766-4bae-bcc7-a6a69f2876f1" (UID: "c1c37cee-a766-4bae-bcc7-a6a69f2876f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.439400 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c1c37cee-a766-4bae-bcc7-a6a69f2876f1" (UID: "c1c37cee-a766-4bae-bcc7-a6a69f2876f1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.441400 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c1c37cee-a766-4bae-bcc7-a6a69f2876f1" (UID: "c1c37cee-a766-4bae-bcc7-a6a69f2876f1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.456298 4629 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.456350 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s86p\" (UniqueName: \"kubernetes.io/projected/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-kube-api-access-9s86p\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.456442 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.456458 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.456487 4629 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c37cee-a766-4bae-bcc7-a6a69f2876f1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.479770 4629 generic.go:334] "Generic (PLEG): container finished" podID="c1c37cee-a766-4bae-bcc7-a6a69f2876f1" containerID="5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd" exitCode=0 Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.479808 4629 generic.go:334] "Generic (PLEG): container finished" podID="c1c37cee-a766-4bae-bcc7-a6a69f2876f1" containerID="de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141" exitCode=143 Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.479869 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1c37cee-a766-4bae-bcc7-a6a69f2876f1","Type":"ContainerDied","Data":"5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd"} Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.479901 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1c37cee-a766-4bae-bcc7-a6a69f2876f1","Type":"ContainerDied","Data":"de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141"} Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.479914 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1c37cee-a766-4bae-bcc7-a6a69f2876f1","Type":"ContainerDied","Data":"60e04aa43945fd26c655a186a90c722a866eba54c3700ed0c6ffae03fa38ca8a"} Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.479930 4629 scope.go:117] "RemoveContainer" containerID="5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.480067 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.491909 4629 generic.go:334] "Generic (PLEG): container finished" podID="cef5a8b9-b204-44a9-92fd-5bd1405895dd" containerID="fc5af8726f2434e79defccb71eecf9c7055a8afef5488f5f6077a554ef738875" exitCode=0 Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.492111 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29424061-mzwc9" event={"ID":"cef5a8b9-b204-44a9-92fd-5bd1405895dd","Type":"ContainerDied","Data":"fc5af8726f2434e79defccb71eecf9c7055a8afef5488f5f6077a554ef738875"} Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.504876 4629 generic.go:334] "Generic (PLEG): container finished" podID="7be44442-eb08-4028-87d9-96decf3215b9" containerID="dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9" exitCode=143 Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.505080 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7be44442-eb08-4028-87d9-96decf3215b9","Type":"ContainerDied","Data":"dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9"} Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.523014 4629 scope.go:117] "RemoveContainer" containerID="de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.523601 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9568824-b16c-4084-90c7-dd6fe60a866a","Type":"ContainerStarted","Data":"38e61685749446c1be5665d06a13ac87a8346d5cc1c4a8e89f8f06053a919157"} Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.524917 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.571052 4629 scope.go:117] "RemoveContainer" containerID="5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd" Dec 11 09:01:04 crc kubenswrapper[4629]: E1211 09:01:04.571779 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd\": container with ID starting with 5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd not found: ID does not exist" containerID="5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.571828 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd"} err="failed to get container status \"5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd\": rpc error: code = NotFound desc = could not find container \"5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd\": container with ID starting with 5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd not found: ID does not exist" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.571878 4629 scope.go:117] "RemoveContainer" containerID="de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141" Dec 11 09:01:04 crc kubenswrapper[4629]: E1211 09:01:04.572143 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141\": container with ID starting with de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141 not found: ID does not exist" containerID="de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.572171 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141"} err="failed to get container status \"de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141\": rpc error: code = NotFound desc = could not find container \"de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141\": container with ID starting with de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141 not found: ID does not exist" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.572189 4629 scope.go:117] "RemoveContainer" containerID="5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.572370 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd"} err="failed to get container status \"5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd\": rpc error: code = NotFound desc = could not find container \"5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd\": container with ID starting with 5e3815b330b5fa61b83b660f374b7dd229fb3b670e2f757dce227aa110e0cbfd not found: ID does not exist" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.572393 4629 scope.go:117] "RemoveContainer" containerID="de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.572574 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141"} err="failed to get container status \"de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141\": rpc error: code = NotFound desc = could not find container \"de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141\": container with ID starting with de106e4c0c28647962f1af1453575a1cab75ef7349667d3bdd0a81b7ed191141 not found: ID does not exist" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.585578 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.623241 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.627644 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 11 09:01:04 crc kubenswrapper[4629]: E1211 09:01:04.628126 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de77a96-4287-46cc-8004-04f26169a35d" containerName="nova-manage" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.628148 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de77a96-4287-46cc-8004-04f26169a35d" containerName="nova-manage" Dec 11 09:01:04 crc kubenswrapper[4629]: E1211 09:01:04.628161 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1c37cee-a766-4bae-bcc7-a6a69f2876f1" containerName="nova-api-log" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.628168 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1c37cee-a766-4bae-bcc7-a6a69f2876f1" containerName="nova-api-log" Dec 11 09:01:04 crc kubenswrapper[4629]: E1211 09:01:04.628180 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1c37cee-a766-4bae-bcc7-a6a69f2876f1" containerName="nova-api-api" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.628188 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1c37cee-a766-4bae-bcc7-a6a69f2876f1" containerName="nova-api-api" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.628409 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="2de77a96-4287-46cc-8004-04f26169a35d" containerName="nova-manage" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.628434 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1c37cee-a766-4bae-bcc7-a6a69f2876f1" containerName="nova-api-api" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.628445 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1c37cee-a766-4bae-bcc7-a6a69f2876f1" containerName="nova-api-log" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.629449 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.634480 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.634730 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.634905 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.638603 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.647315 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.774386034 podStartE2EDuration="7.647291655s" podCreationTimestamp="2025-12-11 09:00:57 +0000 UTC" firstStartedPulling="2025-12-11 09:00:58.441473567 +0000 UTC m=+1446.506889175" lastFinishedPulling="2025-12-11 09:01:03.314379188 +0000 UTC m=+1451.379794796" observedRunningTime="2025-12-11 09:01:04.593683371 +0000 UTC m=+1452.659098979" watchObservedRunningTime="2025-12-11 09:01:04.647291655 +0000 UTC m=+1452.712707273" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.761441 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf7hj\" (UniqueName: \"kubernetes.io/projected/12842a16-5854-4f02-932e-24c306f4a07b-kube-api-access-wf7hj\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.761503 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12842a16-5854-4f02-932e-24c306f4a07b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.761591 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12842a16-5854-4f02-932e-24c306f4a07b-public-tls-certs\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.761677 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12842a16-5854-4f02-932e-24c306f4a07b-config-data\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.761731 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12842a16-5854-4f02-932e-24c306f4a07b-logs\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.761788 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12842a16-5854-4f02-932e-24c306f4a07b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.864087 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12842a16-5854-4f02-932e-24c306f4a07b-config-data\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.864156 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12842a16-5854-4f02-932e-24c306f4a07b-logs\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.864205 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12842a16-5854-4f02-932e-24c306f4a07b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.864257 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf7hj\" (UniqueName: \"kubernetes.io/projected/12842a16-5854-4f02-932e-24c306f4a07b-kube-api-access-wf7hj\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.864295 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12842a16-5854-4f02-932e-24c306f4a07b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.864354 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12842a16-5854-4f02-932e-24c306f4a07b-public-tls-certs\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.864690 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12842a16-5854-4f02-932e-24c306f4a07b-logs\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.870203 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12842a16-5854-4f02-932e-24c306f4a07b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.870794 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12842a16-5854-4f02-932e-24c306f4a07b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.873400 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12842a16-5854-4f02-932e-24c306f4a07b-config-data\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.885943 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12842a16-5854-4f02-932e-24c306f4a07b-public-tls-certs\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.889517 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf7hj\" (UniqueName: \"kubernetes.io/projected/12842a16-5854-4f02-932e-24c306f4a07b-kube-api-access-wf7hj\") pod \"nova-api-0\" (UID: \"12842a16-5854-4f02-932e-24c306f4a07b\") " pod="openstack/nova-api-0" Dec 11 09:01:04 crc kubenswrapper[4629]: I1211 09:01:04.963900 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 09:01:05 crc kubenswrapper[4629]: E1211 09:01:05.274001 4629 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 528aad845f2a7b47b2a5bc924e7bdfb69195902a283f2d14900cfb11c4ec2a0b is running failed: container process not found" containerID="528aad845f2a7b47b2a5bc924e7bdfb69195902a283f2d14900cfb11c4ec2a0b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 11 09:01:05 crc kubenswrapper[4629]: E1211 09:01:05.277540 4629 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 528aad845f2a7b47b2a5bc924e7bdfb69195902a283f2d14900cfb11c4ec2a0b is running failed: container process not found" containerID="528aad845f2a7b47b2a5bc924e7bdfb69195902a283f2d14900cfb11c4ec2a0b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 11 09:01:05 crc kubenswrapper[4629]: E1211 09:01:05.280028 4629 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 528aad845f2a7b47b2a5bc924e7bdfb69195902a283f2d14900cfb11c4ec2a0b is running failed: container process not found" containerID="528aad845f2a7b47b2a5bc924e7bdfb69195902a283f2d14900cfb11c4ec2a0b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 11 09:01:05 crc kubenswrapper[4629]: E1211 09:01:05.280188 4629 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 528aad845f2a7b47b2a5bc924e7bdfb69195902a283f2d14900cfb11c4ec2a0b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="d4c82eb1-f999-45b7-82af-58e3b56a2e0a" containerName="nova-scheduler-scheduler" Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.513351 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.539932 4629 generic.go:334] "Generic (PLEG): container finished" podID="d4c82eb1-f999-45b7-82af-58e3b56a2e0a" containerID="528aad845f2a7b47b2a5bc924e7bdfb69195902a283f2d14900cfb11c4ec2a0b" exitCode=0 Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.540010 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d4c82eb1-f999-45b7-82af-58e3b56a2e0a","Type":"ContainerDied","Data":"528aad845f2a7b47b2a5bc924e7bdfb69195902a283f2d14900cfb11c4ec2a0b"} Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.544776 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12842a16-5854-4f02-932e-24c306f4a07b","Type":"ContainerStarted","Data":"a221290ca8b2b92f2436e774a5fa75ba4f314e008ce88503f6f4bb162cfae0f4"} Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.551618 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.576882 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc22j\" (UniqueName: \"kubernetes.io/projected/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-kube-api-access-zc22j\") pod \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\" (UID: \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\") " Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.577170 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-config-data\") pod \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\" (UID: \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\") " Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.577258 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-combined-ca-bundle\") pod \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\" (UID: \"d4c82eb1-f999-45b7-82af-58e3b56a2e0a\") " Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.584044 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-kube-api-access-zc22j" (OuterVolumeSpecName: "kube-api-access-zc22j") pod "d4c82eb1-f999-45b7-82af-58e3b56a2e0a" (UID: "d4c82eb1-f999-45b7-82af-58e3b56a2e0a"). InnerVolumeSpecName "kube-api-access-zc22j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.619731 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-config-data" (OuterVolumeSpecName: "config-data") pod "d4c82eb1-f999-45b7-82af-58e3b56a2e0a" (UID: "d4c82eb1-f999-45b7-82af-58e3b56a2e0a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.639590 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4c82eb1-f999-45b7-82af-58e3b56a2e0a" (UID: "d4c82eb1-f999-45b7-82af-58e3b56a2e0a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.692121 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.692155 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.692166 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc22j\" (UniqueName: \"kubernetes.io/projected/d4c82eb1-f999-45b7-82af-58e3b56a2e0a-kube-api-access-zc22j\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:05 crc kubenswrapper[4629]: I1211 09:01:05.977375 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.007712 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-combined-ca-bundle\") pod \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.008017 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-config-data\") pod \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.008204 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-fernet-keys\") pod \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.008374 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rplg5\" (UniqueName: \"kubernetes.io/projected/cef5a8b9-b204-44a9-92fd-5bd1405895dd-kube-api-access-rplg5\") pod \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\" (UID: \"cef5a8b9-b204-44a9-92fd-5bd1405895dd\") " Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.012322 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cef5a8b9-b204-44a9-92fd-5bd1405895dd-kube-api-access-rplg5" (OuterVolumeSpecName: "kube-api-access-rplg5") pod "cef5a8b9-b204-44a9-92fd-5bd1405895dd" (UID: "cef5a8b9-b204-44a9-92fd-5bd1405895dd"). InnerVolumeSpecName "kube-api-access-rplg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.024452 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "cef5a8b9-b204-44a9-92fd-5bd1405895dd" (UID: "cef5a8b9-b204-44a9-92fd-5bd1405895dd"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.102108 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cef5a8b9-b204-44a9-92fd-5bd1405895dd" (UID: "cef5a8b9-b204-44a9-92fd-5bd1405895dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.111246 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.111286 4629 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.111299 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rplg5\" (UniqueName: \"kubernetes.io/projected/cef5a8b9-b204-44a9-92fd-5bd1405895dd-kube-api-access-rplg5\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.130976 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-config-data" (OuterVolumeSpecName: "config-data") pod "cef5a8b9-b204-44a9-92fd-5bd1405895dd" (UID: "cef5a8b9-b204-44a9-92fd-5bd1405895dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.210770 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1c37cee-a766-4bae-bcc7-a6a69f2876f1" path="/var/lib/kubelet/pods/c1c37cee-a766-4bae-bcc7-a6a69f2876f1/volumes" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.212800 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cef5a8b9-b204-44a9-92fd-5bd1405895dd-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.581685 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29424061-mzwc9" event={"ID":"cef5a8b9-b204-44a9-92fd-5bd1405895dd","Type":"ContainerDied","Data":"dc9f7255ccae1694425834ed1bb9beabe18e586dee92ff0006f265d32c010c70"} Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.582178 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc9f7255ccae1694425834ed1bb9beabe18e586dee92ff0006f265d32c010c70" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.581714 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29424061-mzwc9" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.584238 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d4c82eb1-f999-45b7-82af-58e3b56a2e0a","Type":"ContainerDied","Data":"62722dd553f0f93dc04a5f464d231d41522dd12e4c2765f260f0264f09bee538"} Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.584344 4629 scope.go:117] "RemoveContainer" containerID="528aad845f2a7b47b2a5bc924e7bdfb69195902a283f2d14900cfb11c4ec2a0b" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.584527 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.590268 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12842a16-5854-4f02-932e-24c306f4a07b","Type":"ContainerStarted","Data":"8ec39bd72f6c16853a72d0a45d10cc99eeaa28f2e0a3d42e19b85b8d9c1ae320"} Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.590321 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12842a16-5854-4f02-932e-24c306f4a07b","Type":"ContainerStarted","Data":"4d96a1619491f7a016f67c900dd422a51293724c105d083281c4bb68134bede9"} Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.614052 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.614031755 podStartE2EDuration="2.614031755s" podCreationTimestamp="2025-12-11 09:01:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:01:06.613143097 +0000 UTC m=+1454.678558705" watchObservedRunningTime="2025-12-11 09:01:06.614031755 +0000 UTC m=+1454.679447363" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.652161 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.672406 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.686354 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.687680 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.697416 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:01:06 crc kubenswrapper[4629]: E1211 09:01:06.698000 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cef5a8b9-b204-44a9-92fd-5bd1405895dd" containerName="keystone-cron" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.698019 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="cef5a8b9-b204-44a9-92fd-5bd1405895dd" containerName="keystone-cron" Dec 11 09:01:06 crc kubenswrapper[4629]: E1211 09:01:06.698047 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c82eb1-f999-45b7-82af-58e3b56a2e0a" containerName="nova-scheduler-scheduler" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.698057 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c82eb1-f999-45b7-82af-58e3b56a2e0a" containerName="nova-scheduler-scheduler" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.698309 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c82eb1-f999-45b7-82af-58e3b56a2e0a" containerName="nova-scheduler-scheduler" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.698323 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="cef5a8b9-b204-44a9-92fd-5bd1405895dd" containerName="keystone-cron" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.699137 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.702679 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.722237 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f99fda5-5b51-4936-a64f-c9e210a1fc64-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7f99fda5-5b51-4936-a64f-c9e210a1fc64\") " pod="openstack/nova-scheduler-0" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.722325 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f99fda5-5b51-4936-a64f-c9e210a1fc64-config-data\") pod \"nova-scheduler-0\" (UID: \"7f99fda5-5b51-4936-a64f-c9e210a1fc64\") " pod="openstack/nova-scheduler-0" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.722395 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rvxb\" (UniqueName: \"kubernetes.io/projected/7f99fda5-5b51-4936-a64f-c9e210a1fc64-kube-api-access-8rvxb\") pod \"nova-scheduler-0\" (UID: \"7f99fda5-5b51-4936-a64f-c9e210a1fc64\") " pod="openstack/nova-scheduler-0" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.731897 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.824399 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f99fda5-5b51-4936-a64f-c9e210a1fc64-config-data\") pod \"nova-scheduler-0\" (UID: \"7f99fda5-5b51-4936-a64f-c9e210a1fc64\") " pod="openstack/nova-scheduler-0" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.824593 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rvxb\" (UniqueName: \"kubernetes.io/projected/7f99fda5-5b51-4936-a64f-c9e210a1fc64-kube-api-access-8rvxb\") pod \"nova-scheduler-0\" (UID: \"7f99fda5-5b51-4936-a64f-c9e210a1fc64\") " pod="openstack/nova-scheduler-0" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.827978 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f99fda5-5b51-4936-a64f-c9e210a1fc64-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7f99fda5-5b51-4936-a64f-c9e210a1fc64\") " pod="openstack/nova-scheduler-0" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.833087 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f99fda5-5b51-4936-a64f-c9e210a1fc64-config-data\") pod \"nova-scheduler-0\" (UID: \"7f99fda5-5b51-4936-a64f-c9e210a1fc64\") " pod="openstack/nova-scheduler-0" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.839246 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f99fda5-5b51-4936-a64f-c9e210a1fc64-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7f99fda5-5b51-4936-a64f-c9e210a1fc64\") " pod="openstack/nova-scheduler-0" Dec 11 09:01:06 crc kubenswrapper[4629]: I1211 09:01:06.842052 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rvxb\" (UniqueName: \"kubernetes.io/projected/7f99fda5-5b51-4936-a64f-c9e210a1fc64-kube-api-access-8rvxb\") pod \"nova-scheduler-0\" (UID: \"7f99fda5-5b51-4936-a64f-c9e210a1fc64\") " pod="openstack/nova-scheduler-0" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.034634 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.115610 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="7be44442-eb08-4028-87d9-96decf3215b9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.178:8775/\": read tcp 10.217.0.2:58852->10.217.0.178:8775: read: connection reset by peer" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.116077 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="7be44442-eb08-4028-87d9-96decf3215b9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.178:8775/\": read tcp 10.217.0.2:58868->10.217.0.178:8775: read: connection reset by peer" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.498918 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.528149 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.621264 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7f99fda5-5b51-4936-a64f-c9e210a1fc64","Type":"ContainerStarted","Data":"7d97b9ada5f9f79cb23f1099a20a79b73fb8b41a42341966337541aa2fbc5f73"} Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.627324 4629 generic.go:334] "Generic (PLEG): container finished" podID="7be44442-eb08-4028-87d9-96decf3215b9" containerID="fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a" exitCode=0 Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.627626 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.627647 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7be44442-eb08-4028-87d9-96decf3215b9","Type":"ContainerDied","Data":"fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a"} Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.627916 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7be44442-eb08-4028-87d9-96decf3215b9","Type":"ContainerDied","Data":"d432fbe5e4497cf2ad6644b72d2797da35d166bf3cc76328a0e6323d0de4d2cf"} Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.627949 4629 scope.go:117] "RemoveContainer" containerID="fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.692192 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-nova-metadata-tls-certs\") pod \"7be44442-eb08-4028-87d9-96decf3215b9\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.700253 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-config-data\") pod \"7be44442-eb08-4028-87d9-96decf3215b9\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.700415 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lv5m\" (UniqueName: \"kubernetes.io/projected/7be44442-eb08-4028-87d9-96decf3215b9-kube-api-access-7lv5m\") pod \"7be44442-eb08-4028-87d9-96decf3215b9\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.700492 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-combined-ca-bundle\") pod \"7be44442-eb08-4028-87d9-96decf3215b9\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.700519 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7be44442-eb08-4028-87d9-96decf3215b9-logs\") pod \"7be44442-eb08-4028-87d9-96decf3215b9\" (UID: \"7be44442-eb08-4028-87d9-96decf3215b9\") " Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.703009 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7be44442-eb08-4028-87d9-96decf3215b9-logs" (OuterVolumeSpecName: "logs") pod "7be44442-eb08-4028-87d9-96decf3215b9" (UID: "7be44442-eb08-4028-87d9-96decf3215b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.709490 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7be44442-eb08-4028-87d9-96decf3215b9-kube-api-access-7lv5m" (OuterVolumeSpecName: "kube-api-access-7lv5m") pod "7be44442-eb08-4028-87d9-96decf3215b9" (UID: "7be44442-eb08-4028-87d9-96decf3215b9"). InnerVolumeSpecName "kube-api-access-7lv5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.754167 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7be44442-eb08-4028-87d9-96decf3215b9" (UID: "7be44442-eb08-4028-87d9-96decf3215b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.766192 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4xpfw" podUID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" containerName="registry-server" probeResult="failure" output=< Dec 11 09:01:07 crc kubenswrapper[4629]: timeout: failed to connect service ":50051" within 1s Dec 11 09:01:07 crc kubenswrapper[4629]: > Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.775071 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "7be44442-eb08-4028-87d9-96decf3215b9" (UID: "7be44442-eb08-4028-87d9-96decf3215b9"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.805713 4629 scope.go:117] "RemoveContainer" containerID="dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.811773 4629 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.811810 4629 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7be44442-eb08-4028-87d9-96decf3215b9-logs\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.811821 4629 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.811832 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lv5m\" (UniqueName: \"kubernetes.io/projected/7be44442-eb08-4028-87d9-96decf3215b9-kube-api-access-7lv5m\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.879783 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-config-data" (OuterVolumeSpecName: "config-data") pod "7be44442-eb08-4028-87d9-96decf3215b9" (UID: "7be44442-eb08-4028-87d9-96decf3215b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.879907 4629 scope.go:117] "RemoveContainer" containerID="fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a" Dec 11 09:01:07 crc kubenswrapper[4629]: E1211 09:01:07.880377 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a\": container with ID starting with fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a not found: ID does not exist" containerID="fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.880421 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a"} err="failed to get container status \"fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a\": rpc error: code = NotFound desc = could not find container \"fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a\": container with ID starting with fdb59970f27aa96a71cb94ede0df1ac3c2d6e5c18f9f473afe3db26d8ffad05a not found: ID does not exist" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.880454 4629 scope.go:117] "RemoveContainer" containerID="dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9" Dec 11 09:01:07 crc kubenswrapper[4629]: E1211 09:01:07.884944 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9\": container with ID starting with dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9 not found: ID does not exist" containerID="dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.885049 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9"} err="failed to get container status \"dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9\": rpc error: code = NotFound desc = could not find container \"dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9\": container with ID starting with dcb5076a950900e7f9b3a223f5021355362695650db1fa8d45929b99124a7da9 not found: ID does not exist" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.914740 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7be44442-eb08-4028-87d9-96decf3215b9-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.970299 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.984887 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.996240 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:01:07 crc kubenswrapper[4629]: E1211 09:01:07.996874 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7be44442-eb08-4028-87d9-96decf3215b9" containerName="nova-metadata-metadata" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.996954 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="7be44442-eb08-4028-87d9-96decf3215b9" containerName="nova-metadata-metadata" Dec 11 09:01:07 crc kubenswrapper[4629]: E1211 09:01:07.997068 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7be44442-eb08-4028-87d9-96decf3215b9" containerName="nova-metadata-log" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.997147 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="7be44442-eb08-4028-87d9-96decf3215b9" containerName="nova-metadata-log" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.997455 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="7be44442-eb08-4028-87d9-96decf3215b9" containerName="nova-metadata-metadata" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.997622 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="7be44442-eb08-4028-87d9-96decf3215b9" containerName="nova-metadata-log" Dec 11 09:01:07 crc kubenswrapper[4629]: I1211 09:01:07.998824 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.005669 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.005988 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.019422 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.118642 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff06114c-4d4f-4868-a19c-0598aee0ab52-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.119012 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff06114c-4d4f-4868-a19c-0598aee0ab52-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.119157 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff06114c-4d4f-4868-a19c-0598aee0ab52-config-data\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.119271 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8vrv\" (UniqueName: \"kubernetes.io/projected/ff06114c-4d4f-4868-a19c-0598aee0ab52-kube-api-access-b8vrv\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.119524 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff06114c-4d4f-4868-a19c-0598aee0ab52-logs\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.211294 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7be44442-eb08-4028-87d9-96decf3215b9" path="/var/lib/kubelet/pods/7be44442-eb08-4028-87d9-96decf3215b9/volumes" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.212402 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4c82eb1-f999-45b7-82af-58e3b56a2e0a" path="/var/lib/kubelet/pods/d4c82eb1-f999-45b7-82af-58e3b56a2e0a/volumes" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.220914 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff06114c-4d4f-4868-a19c-0598aee0ab52-logs\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.221031 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff06114c-4d4f-4868-a19c-0598aee0ab52-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.221160 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff06114c-4d4f-4868-a19c-0598aee0ab52-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.221229 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff06114c-4d4f-4868-a19c-0598aee0ab52-config-data\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.221261 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8vrv\" (UniqueName: \"kubernetes.io/projected/ff06114c-4d4f-4868-a19c-0598aee0ab52-kube-api-access-b8vrv\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.222121 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff06114c-4d4f-4868-a19c-0598aee0ab52-logs\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.227962 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff06114c-4d4f-4868-a19c-0598aee0ab52-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.229620 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff06114c-4d4f-4868-a19c-0598aee0ab52-config-data\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.230057 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff06114c-4d4f-4868-a19c-0598aee0ab52-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.237418 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8vrv\" (UniqueName: \"kubernetes.io/projected/ff06114c-4d4f-4868-a19c-0598aee0ab52-kube-api-access-b8vrv\") pod \"nova-metadata-0\" (UID: \"ff06114c-4d4f-4868-a19c-0598aee0ab52\") " pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.321362 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.710451 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7f99fda5-5b51-4936-a64f-c9e210a1fc64","Type":"ContainerStarted","Data":"be409b4992435a9a656387d813199fbc5c8931854e977c21e6bc31e8b5acff15"} Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.729342 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.729321831 podStartE2EDuration="2.729321831s" podCreationTimestamp="2025-12-11 09:01:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:01:08.72771866 +0000 UTC m=+1456.793134288" watchObservedRunningTime="2025-12-11 09:01:08.729321831 +0000 UTC m=+1456.794737439" Dec 11 09:01:08 crc kubenswrapper[4629]: I1211 09:01:08.811177 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 09:01:08 crc kubenswrapper[4629]: W1211 09:01:08.818046 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff06114c_4d4f_4868_a19c_0598aee0ab52.slice/crio-34f2d6aa6e885e2bc4511dc2a3bcba23230abfa3c87ab58d4b25803d4d52d049 WatchSource:0}: Error finding container 34f2d6aa6e885e2bc4511dc2a3bcba23230abfa3c87ab58d4b25803d4d52d049: Status 404 returned error can't find the container with id 34f2d6aa6e885e2bc4511dc2a3bcba23230abfa3c87ab58d4b25803d4d52d049 Dec 11 09:01:09 crc kubenswrapper[4629]: I1211 09:01:09.721660 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff06114c-4d4f-4868-a19c-0598aee0ab52","Type":"ContainerStarted","Data":"dffd53aac31ee75cac6b572f0ed046146ca8f0a72baa5638cb653c14364bfd88"} Dec 11 09:01:09 crc kubenswrapper[4629]: I1211 09:01:09.722072 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff06114c-4d4f-4868-a19c-0598aee0ab52","Type":"ContainerStarted","Data":"75f0541d557f441bf4aee720e8bf3c2bd458a315ff86656398801424a89592de"} Dec 11 09:01:09 crc kubenswrapper[4629]: I1211 09:01:09.722088 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff06114c-4d4f-4868-a19c-0598aee0ab52","Type":"ContainerStarted","Data":"34f2d6aa6e885e2bc4511dc2a3bcba23230abfa3c87ab58d4b25803d4d52d049"} Dec 11 09:01:09 crc kubenswrapper[4629]: I1211 09:01:09.786771 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.786752306 podStartE2EDuration="2.786752306s" podCreationTimestamp="2025-12-11 09:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:01:09.78656825 +0000 UTC m=+1457.851983858" watchObservedRunningTime="2025-12-11 09:01:09.786752306 +0000 UTC m=+1457.852167914" Dec 11 09:01:12 crc kubenswrapper[4629]: I1211 09:01:12.035480 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 11 09:01:13 crc kubenswrapper[4629]: I1211 09:01:13.328116 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 09:01:13 crc kubenswrapper[4629]: I1211 09:01:13.329059 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 09:01:14 crc kubenswrapper[4629]: I1211 09:01:14.965166 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 09:01:14 crc kubenswrapper[4629]: I1211 09:01:14.965217 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 09:01:15 crc kubenswrapper[4629]: I1211 09:01:15.978029 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="12842a16-5854-4f02-932e-24c306f4a07b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.186:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 09:01:15 crc kubenswrapper[4629]: I1211 09:01:15.978027 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="12842a16-5854-4f02-932e-24c306f4a07b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.186:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 09:01:16 crc kubenswrapper[4629]: I1211 09:01:16.744613 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:01:16 crc kubenswrapper[4629]: I1211 09:01:16.801100 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:01:17 crc kubenswrapper[4629]: I1211 09:01:17.036074 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 11 09:01:17 crc kubenswrapper[4629]: I1211 09:01:17.072407 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 11 09:01:17 crc kubenswrapper[4629]: I1211 09:01:17.693202 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4xpfw"] Dec 11 09:01:17 crc kubenswrapper[4629]: I1211 09:01:17.811941 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4xpfw" podUID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" containerName="registry-server" containerID="cri-o://650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f" gracePeriod=2 Dec 11 09:01:17 crc kubenswrapper[4629]: I1211 09:01:17.841831 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.285268 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.327764 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.327817 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.389026 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed4090d-b080-4da4-bb8d-bffd8a1680b2-catalog-content\") pod \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\" (UID: \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\") " Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.389306 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qj57\" (UniqueName: \"kubernetes.io/projected/aed4090d-b080-4da4-bb8d-bffd8a1680b2-kube-api-access-8qj57\") pod \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\" (UID: \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\") " Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.389340 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed4090d-b080-4da4-bb8d-bffd8a1680b2-utilities\") pod \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\" (UID: \"aed4090d-b080-4da4-bb8d-bffd8a1680b2\") " Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.392255 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aed4090d-b080-4da4-bb8d-bffd8a1680b2-utilities" (OuterVolumeSpecName: "utilities") pod "aed4090d-b080-4da4-bb8d-bffd8a1680b2" (UID: "aed4090d-b080-4da4-bb8d-bffd8a1680b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.398134 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed4090d-b080-4da4-bb8d-bffd8a1680b2-kube-api-access-8qj57" (OuterVolumeSpecName: "kube-api-access-8qj57") pod "aed4090d-b080-4da4-bb8d-bffd8a1680b2" (UID: "aed4090d-b080-4da4-bb8d-bffd8a1680b2"). InnerVolumeSpecName "kube-api-access-8qj57". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.492366 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qj57\" (UniqueName: \"kubernetes.io/projected/aed4090d-b080-4da4-bb8d-bffd8a1680b2-kube-api-access-8qj57\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.492410 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed4090d-b080-4da4-bb8d-bffd8a1680b2-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.537952 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aed4090d-b080-4da4-bb8d-bffd8a1680b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aed4090d-b080-4da4-bb8d-bffd8a1680b2" (UID: "aed4090d-b080-4da4-bb8d-bffd8a1680b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.614787 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed4090d-b080-4da4-bb8d-bffd8a1680b2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.823229 4629 generic.go:334] "Generic (PLEG): container finished" podID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" containerID="650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f" exitCode=0 Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.823275 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xpfw" event={"ID":"aed4090d-b080-4da4-bb8d-bffd8a1680b2","Type":"ContainerDied","Data":"650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f"} Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.823303 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xpfw" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.823353 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xpfw" event={"ID":"aed4090d-b080-4da4-bb8d-bffd8a1680b2","Type":"ContainerDied","Data":"e20cf1a3746663723eed2738908e1dc7491d20a6768b44caff04b5e8d3dff9c2"} Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.823384 4629 scope.go:117] "RemoveContainer" containerID="650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.856418 4629 scope.go:117] "RemoveContainer" containerID="87df4abdc5f28de3fd89dba33399049ee90091f7be28376c7edec2b6879edf14" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.890174 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4xpfw"] Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.893488 4629 scope.go:117] "RemoveContainer" containerID="bc907957c33da7c62363adf0334d5290dc1e2781513678ef7a444c18a0e12434" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.920255 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4xpfw"] Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.969089 4629 scope.go:117] "RemoveContainer" containerID="650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f" Dec 11 09:01:18 crc kubenswrapper[4629]: E1211 09:01:18.969928 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f\": container with ID starting with 650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f not found: ID does not exist" containerID="650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.969966 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f"} err="failed to get container status \"650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f\": rpc error: code = NotFound desc = could not find container \"650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f\": container with ID starting with 650aa586fd3b914760c586feb8613d4a82075b12811447862dd88132350f8a7f not found: ID does not exist" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.969992 4629 scope.go:117] "RemoveContainer" containerID="87df4abdc5f28de3fd89dba33399049ee90091f7be28376c7edec2b6879edf14" Dec 11 09:01:18 crc kubenswrapper[4629]: E1211 09:01:18.970500 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87df4abdc5f28de3fd89dba33399049ee90091f7be28376c7edec2b6879edf14\": container with ID starting with 87df4abdc5f28de3fd89dba33399049ee90091f7be28376c7edec2b6879edf14 not found: ID does not exist" containerID="87df4abdc5f28de3fd89dba33399049ee90091f7be28376c7edec2b6879edf14" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.970533 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87df4abdc5f28de3fd89dba33399049ee90091f7be28376c7edec2b6879edf14"} err="failed to get container status \"87df4abdc5f28de3fd89dba33399049ee90091f7be28376c7edec2b6879edf14\": rpc error: code = NotFound desc = could not find container \"87df4abdc5f28de3fd89dba33399049ee90091f7be28376c7edec2b6879edf14\": container with ID starting with 87df4abdc5f28de3fd89dba33399049ee90091f7be28376c7edec2b6879edf14 not found: ID does not exist" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.970551 4629 scope.go:117] "RemoveContainer" containerID="bc907957c33da7c62363adf0334d5290dc1e2781513678ef7a444c18a0e12434" Dec 11 09:01:18 crc kubenswrapper[4629]: E1211 09:01:18.971196 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc907957c33da7c62363adf0334d5290dc1e2781513678ef7a444c18a0e12434\": container with ID starting with bc907957c33da7c62363adf0334d5290dc1e2781513678ef7a444c18a0e12434 not found: ID does not exist" containerID="bc907957c33da7c62363adf0334d5290dc1e2781513678ef7a444c18a0e12434" Dec 11 09:01:18 crc kubenswrapper[4629]: I1211 09:01:18.971225 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc907957c33da7c62363adf0334d5290dc1e2781513678ef7a444c18a0e12434"} err="failed to get container status \"bc907957c33da7c62363adf0334d5290dc1e2781513678ef7a444c18a0e12434\": rpc error: code = NotFound desc = could not find container \"bc907957c33da7c62363adf0334d5290dc1e2781513678ef7a444c18a0e12434\": container with ID starting with bc907957c33da7c62363adf0334d5290dc1e2781513678ef7a444c18a0e12434 not found: ID does not exist" Dec 11 09:01:19 crc kubenswrapper[4629]: I1211 09:01:19.410111 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ff06114c-4d4f-4868-a19c-0598aee0ab52" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.188:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 09:01:19 crc kubenswrapper[4629]: I1211 09:01:19.414137 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ff06114c-4d4f-4868-a19c-0598aee0ab52" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.188:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 09:01:20 crc kubenswrapper[4629]: I1211 09:01:20.210565 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" path="/var/lib/kubelet/pods/aed4090d-b080-4da4-bb8d-bffd8a1680b2/volumes" Dec 11 09:01:24 crc kubenswrapper[4629]: I1211 09:01:24.985456 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 11 09:01:24 crc kubenswrapper[4629]: I1211 09:01:24.986400 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 11 09:01:24 crc kubenswrapper[4629]: I1211 09:01:24.987531 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 11 09:01:24 crc kubenswrapper[4629]: I1211 09:01:24.997558 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 11 09:01:25 crc kubenswrapper[4629]: I1211 09:01:25.880568 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 11 09:01:25 crc kubenswrapper[4629]: I1211 09:01:25.887687 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 11 09:01:27 crc kubenswrapper[4629]: I1211 09:01:27.923229 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 11 09:01:28 crc kubenswrapper[4629]: I1211 09:01:28.333618 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 11 09:01:28 crc kubenswrapper[4629]: I1211 09:01:28.338703 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 11 09:01:28 crc kubenswrapper[4629]: I1211 09:01:28.339108 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 11 09:01:28 crc kubenswrapper[4629]: I1211 09:01:28.912816 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 11 09:01:37 crc kubenswrapper[4629]: I1211 09:01:37.260411 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 09:01:38 crc kubenswrapper[4629]: I1211 09:01:38.904954 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 09:01:41 crc kubenswrapper[4629]: I1211 09:01:41.785162 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="1a670dab-0f15-4572-b2a8-36ec2330adb2" containerName="rabbitmq" containerID="cri-o://9c73bc96a9bd46a33bc825a170d9b9449a4f62874ff606906eaa78654b7bdebb" gracePeriod=604796 Dec 11 09:01:43 crc kubenswrapper[4629]: I1211 09:01:43.682255 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="0c633337-5e1d-438b-8b18-0acb62ec2da0" containerName="rabbitmq" containerID="cri-o://d414f34d90511b7735512304b7e2e0d05e454ca4546787b2f1fd13a79aaddd76" gracePeriod=604796 Dec 11 09:01:47 crc kubenswrapper[4629]: I1211 09:01:47.542198 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="1a670dab-0f15-4572-b2a8-36ec2330adb2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 11 09:01:47 crc kubenswrapper[4629]: I1211 09:01:47.669078 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="0c633337-5e1d-438b-8b18-0acb62ec2da0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.127171 4629 generic.go:334] "Generic (PLEG): container finished" podID="1a670dab-0f15-4572-b2a8-36ec2330adb2" containerID="9c73bc96a9bd46a33bc825a170d9b9449a4f62874ff606906eaa78654b7bdebb" exitCode=0 Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.127362 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1a670dab-0f15-4572-b2a8-36ec2330adb2","Type":"ContainerDied","Data":"9c73bc96a9bd46a33bc825a170d9b9449a4f62874ff606906eaa78654b7bdebb"} Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.326446 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.497932 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-tls\") pod \"1a670dab-0f15-4572-b2a8-36ec2330adb2\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.498333 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1a670dab-0f15-4572-b2a8-36ec2330adb2-erlang-cookie-secret\") pod \"1a670dab-0f15-4572-b2a8-36ec2330adb2\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.498410 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jddpj\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-kube-api-access-jddpj\") pod \"1a670dab-0f15-4572-b2a8-36ec2330adb2\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.498469 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-confd\") pod \"1a670dab-0f15-4572-b2a8-36ec2330adb2\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.498503 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-erlang-cookie\") pod \"1a670dab-0f15-4572-b2a8-36ec2330adb2\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.498534 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1a670dab-0f15-4572-b2a8-36ec2330adb2-pod-info\") pod \"1a670dab-0f15-4572-b2a8-36ec2330adb2\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.498592 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-plugins\") pod \"1a670dab-0f15-4572-b2a8-36ec2330adb2\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.498695 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-server-conf\") pod \"1a670dab-0f15-4572-b2a8-36ec2330adb2\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.498716 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"1a670dab-0f15-4572-b2a8-36ec2330adb2\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.498754 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-plugins-conf\") pod \"1a670dab-0f15-4572-b2a8-36ec2330adb2\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.498808 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-config-data\") pod \"1a670dab-0f15-4572-b2a8-36ec2330adb2\" (UID: \"1a670dab-0f15-4572-b2a8-36ec2330adb2\") " Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.499431 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1a670dab-0f15-4572-b2a8-36ec2330adb2" (UID: "1a670dab-0f15-4572-b2a8-36ec2330adb2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.499509 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1a670dab-0f15-4572-b2a8-36ec2330adb2" (UID: "1a670dab-0f15-4572-b2a8-36ec2330adb2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.500062 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1a670dab-0f15-4572-b2a8-36ec2330adb2" (UID: "1a670dab-0f15-4572-b2a8-36ec2330adb2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.501049 4629 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.501079 4629 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.501098 4629 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.506881 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1a670dab-0f15-4572-b2a8-36ec2330adb2-pod-info" (OuterVolumeSpecName: "pod-info") pod "1a670dab-0f15-4572-b2a8-36ec2330adb2" (UID: "1a670dab-0f15-4572-b2a8-36ec2330adb2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.507013 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a670dab-0f15-4572-b2a8-36ec2330adb2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1a670dab-0f15-4572-b2a8-36ec2330adb2" (UID: "1a670dab-0f15-4572-b2a8-36ec2330adb2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.508578 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "1a670dab-0f15-4572-b2a8-36ec2330adb2" (UID: "1a670dab-0f15-4572-b2a8-36ec2330adb2"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.508741 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1a670dab-0f15-4572-b2a8-36ec2330adb2" (UID: "1a670dab-0f15-4572-b2a8-36ec2330adb2"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.543107 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-kube-api-access-jddpj" (OuterVolumeSpecName: "kube-api-access-jddpj") pod "1a670dab-0f15-4572-b2a8-36ec2330adb2" (UID: "1a670dab-0f15-4572-b2a8-36ec2330adb2"). InnerVolumeSpecName "kube-api-access-jddpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.557592 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-config-data" (OuterVolumeSpecName: "config-data") pod "1a670dab-0f15-4572-b2a8-36ec2330adb2" (UID: "1a670dab-0f15-4572-b2a8-36ec2330adb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.603052 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.603083 4629 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.603097 4629 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1a670dab-0f15-4572-b2a8-36ec2330adb2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.603114 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jddpj\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-kube-api-access-jddpj\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.603126 4629 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1a670dab-0f15-4572-b2a8-36ec2330adb2-pod-info\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.603171 4629 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.632084 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-server-conf" (OuterVolumeSpecName: "server-conf") pod "1a670dab-0f15-4572-b2a8-36ec2330adb2" (UID: "1a670dab-0f15-4572-b2a8-36ec2330adb2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.636322 4629 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.644620 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1a670dab-0f15-4572-b2a8-36ec2330adb2" (UID: "1a670dab-0f15-4572-b2a8-36ec2330adb2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.705250 4629 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1a670dab-0f15-4572-b2a8-36ec2330adb2-server-conf\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.705290 4629 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:48 crc kubenswrapper[4629]: I1211 09:01:48.705302 4629 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1a670dab-0f15-4572-b2a8-36ec2330adb2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.145201 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1a670dab-0f15-4572-b2a8-36ec2330adb2","Type":"ContainerDied","Data":"7e87e1dc0f3024f6b70832ef4d267cce03220feb662ca7da8af51d7557f0ad31"} Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.145286 4629 scope.go:117] "RemoveContainer" containerID="9c73bc96a9bd46a33bc825a170d9b9449a4f62874ff606906eaa78654b7bdebb" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.145442 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.179440 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.181786 4629 scope.go:117] "RemoveContainer" containerID="d3ec5182b5c5afada44fd8ab808ed3ad84b1a4b2fbc5e3bd1a17034f8d925a9a" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.187196 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.238472 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 09:01:49 crc kubenswrapper[4629]: E1211 09:01:49.238974 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" containerName="registry-server" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.239002 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" containerName="registry-server" Dec 11 09:01:49 crc kubenswrapper[4629]: E1211 09:01:49.239027 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" containerName="extract-content" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.239036 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" containerName="extract-content" Dec 11 09:01:49 crc kubenswrapper[4629]: E1211 09:01:49.239067 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a670dab-0f15-4572-b2a8-36ec2330adb2" containerName="setup-container" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.239078 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a670dab-0f15-4572-b2a8-36ec2330adb2" containerName="setup-container" Dec 11 09:01:49 crc kubenswrapper[4629]: E1211 09:01:49.239264 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" containerName="extract-utilities" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.239272 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" containerName="extract-utilities" Dec 11 09:01:49 crc kubenswrapper[4629]: E1211 09:01:49.239297 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a670dab-0f15-4572-b2a8-36ec2330adb2" containerName="rabbitmq" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.239306 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a670dab-0f15-4572-b2a8-36ec2330adb2" containerName="rabbitmq" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.239505 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed4090d-b080-4da4-bb8d-bffd8a1680b2" containerName="registry-server" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.239539 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a670dab-0f15-4572-b2a8-36ec2330adb2" containerName="rabbitmq" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.240809 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.244378 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.244480 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.245197 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.245362 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.245635 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.246689 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-t52c7" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.247611 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.262147 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.421188 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh96s\" (UniqueName: \"kubernetes.io/projected/8c10d7a6-61cb-4dd8-884c-1672193f2d16-kube-api-access-jh96s\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.421596 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8c10d7a6-61cb-4dd8-884c-1672193f2d16-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.421763 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8c10d7a6-61cb-4dd8-884c-1672193f2d16-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.421932 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.422040 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8c10d7a6-61cb-4dd8-884c-1672193f2d16-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.422158 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8c10d7a6-61cb-4dd8-884c-1672193f2d16-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.422295 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8c10d7a6-61cb-4dd8-884c-1672193f2d16-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.422417 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8c10d7a6-61cb-4dd8-884c-1672193f2d16-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.422563 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8c10d7a6-61cb-4dd8-884c-1672193f2d16-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.422673 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c10d7a6-61cb-4dd8-884c-1672193f2d16-config-data\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.422815 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8c10d7a6-61cb-4dd8-884c-1672193f2d16-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.524323 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh96s\" (UniqueName: \"kubernetes.io/projected/8c10d7a6-61cb-4dd8-884c-1672193f2d16-kube-api-access-jh96s\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.524414 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8c10d7a6-61cb-4dd8-884c-1672193f2d16-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.524451 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8c10d7a6-61cb-4dd8-884c-1672193f2d16-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.524478 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.524495 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8c10d7a6-61cb-4dd8-884c-1672193f2d16-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.524513 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8c10d7a6-61cb-4dd8-884c-1672193f2d16-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.524531 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8c10d7a6-61cb-4dd8-884c-1672193f2d16-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.524547 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8c10d7a6-61cb-4dd8-884c-1672193f2d16-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.524568 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8c10d7a6-61cb-4dd8-884c-1672193f2d16-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.524583 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c10d7a6-61cb-4dd8-884c-1672193f2d16-config-data\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.524611 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8c10d7a6-61cb-4dd8-884c-1672193f2d16-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.525801 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8c10d7a6-61cb-4dd8-884c-1672193f2d16-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.527344 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8c10d7a6-61cb-4dd8-884c-1672193f2d16-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.527957 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8c10d7a6-61cb-4dd8-884c-1672193f2d16-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.528496 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8c10d7a6-61cb-4dd8-884c-1672193f2d16-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.529107 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c10d7a6-61cb-4dd8-884c-1672193f2d16-config-data\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.530189 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8c10d7a6-61cb-4dd8-884c-1672193f2d16-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.531489 4629 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.533439 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8c10d7a6-61cb-4dd8-884c-1672193f2d16-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.534490 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8c10d7a6-61cb-4dd8-884c-1672193f2d16-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.555578 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8c10d7a6-61cb-4dd8-884c-1672193f2d16-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.602051 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.603494 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh96s\" (UniqueName: \"kubernetes.io/projected/8c10d7a6-61cb-4dd8-884c-1672193f2d16-kube-api-access-jh96s\") pod \"rabbitmq-server-0\" (UID: \"8c10d7a6-61cb-4dd8-884c-1672193f2d16\") " pod="openstack/rabbitmq-server-0" Dec 11 09:01:49 crc kubenswrapper[4629]: I1211 09:01:49.862132 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.177483 4629 generic.go:334] "Generic (PLEG): container finished" podID="0c633337-5e1d-438b-8b18-0acb62ec2da0" containerID="d414f34d90511b7735512304b7e2e0d05e454ca4546787b2f1fd13a79aaddd76" exitCode=0 Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.177624 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0c633337-5e1d-438b-8b18-0acb62ec2da0","Type":"ContainerDied","Data":"d414f34d90511b7735512304b7e2e0d05e454ca4546787b2f1fd13a79aaddd76"} Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.240203 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a670dab-0f15-4572-b2a8-36ec2330adb2" path="/var/lib/kubelet/pods/1a670dab-0f15-4572-b2a8-36ec2330adb2/volumes" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.316091 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.450185 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-server-conf\") pod \"0c633337-5e1d-438b-8b18-0acb62ec2da0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.450291 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"0c633337-5e1d-438b-8b18-0acb62ec2da0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.450349 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-plugins-conf\") pod \"0c633337-5e1d-438b-8b18-0acb62ec2da0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.450401 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-tls\") pod \"0c633337-5e1d-438b-8b18-0acb62ec2da0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.450454 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c633337-5e1d-438b-8b18-0acb62ec2da0-pod-info\") pod \"0c633337-5e1d-438b-8b18-0acb62ec2da0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.450488 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-config-data\") pod \"0c633337-5e1d-438b-8b18-0acb62ec2da0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.450508 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9zzh\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-kube-api-access-z9zzh\") pod \"0c633337-5e1d-438b-8b18-0acb62ec2da0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.450538 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-erlang-cookie\") pod \"0c633337-5e1d-438b-8b18-0acb62ec2da0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.450558 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-plugins\") pod \"0c633337-5e1d-438b-8b18-0acb62ec2da0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.450577 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c633337-5e1d-438b-8b18-0acb62ec2da0-erlang-cookie-secret\") pod \"0c633337-5e1d-438b-8b18-0acb62ec2da0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.450615 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-confd\") pod \"0c633337-5e1d-438b-8b18-0acb62ec2da0\" (UID: \"0c633337-5e1d-438b-8b18-0acb62ec2da0\") " Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.454386 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0c633337-5e1d-438b-8b18-0acb62ec2da0" (UID: "0c633337-5e1d-438b-8b18-0acb62ec2da0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.455524 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0c633337-5e1d-438b-8b18-0acb62ec2da0" (UID: "0c633337-5e1d-438b-8b18-0acb62ec2da0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.458741 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0c633337-5e1d-438b-8b18-0acb62ec2da0" (UID: "0c633337-5e1d-438b-8b18-0acb62ec2da0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.459537 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.464259 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0c633337-5e1d-438b-8b18-0acb62ec2da0" (UID: "0c633337-5e1d-438b-8b18-0acb62ec2da0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.467108 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "0c633337-5e1d-438b-8b18-0acb62ec2da0" (UID: "0c633337-5e1d-438b-8b18-0acb62ec2da0"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.472090 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0c633337-5e1d-438b-8b18-0acb62ec2da0-pod-info" (OuterVolumeSpecName: "pod-info") pod "0c633337-5e1d-438b-8b18-0acb62ec2da0" (UID: "0c633337-5e1d-438b-8b18-0acb62ec2da0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.478376 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-kube-api-access-z9zzh" (OuterVolumeSpecName: "kube-api-access-z9zzh") pod "0c633337-5e1d-438b-8b18-0acb62ec2da0" (UID: "0c633337-5e1d-438b-8b18-0acb62ec2da0"). InnerVolumeSpecName "kube-api-access-z9zzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.494476 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c633337-5e1d-438b-8b18-0acb62ec2da0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0c633337-5e1d-438b-8b18-0acb62ec2da0" (UID: "0c633337-5e1d-438b-8b18-0acb62ec2da0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.528202 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-config-data" (OuterVolumeSpecName: "config-data") pod "0c633337-5e1d-438b-8b18-0acb62ec2da0" (UID: "0c633337-5e1d-438b-8b18-0acb62ec2da0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.553547 4629 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.553767 4629 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.553782 4629 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.553791 4629 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c633337-5e1d-438b-8b18-0acb62ec2da0-pod-info\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.553799 4629 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.553808 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9zzh\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-kube-api-access-z9zzh\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.553819 4629 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.553827 4629 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.553836 4629 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c633337-5e1d-438b-8b18-0acb62ec2da0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.558325 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-server-conf" (OuterVolumeSpecName: "server-conf") pod "0c633337-5e1d-438b-8b18-0acb62ec2da0" (UID: "0c633337-5e1d-438b-8b18-0acb62ec2da0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.577609 4629 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.635050 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0c633337-5e1d-438b-8b18-0acb62ec2da0" (UID: "0c633337-5e1d-438b-8b18-0acb62ec2da0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.656045 4629 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c633337-5e1d-438b-8b18-0acb62ec2da0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.656075 4629 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c633337-5e1d-438b-8b18-0acb62ec2da0-server-conf\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:50 crc kubenswrapper[4629]: I1211 09:01:50.656086 4629 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.190498 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.190514 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0c633337-5e1d-438b-8b18-0acb62ec2da0","Type":"ContainerDied","Data":"2b8d148c377e4e26f06811de5ae1ce5d87f41bc23dc8e83dbbf889b9a4847951"} Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.190607 4629 scope.go:117] "RemoveContainer" containerID="d414f34d90511b7735512304b7e2e0d05e454ca4546787b2f1fd13a79aaddd76" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.192647 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8c10d7a6-61cb-4dd8-884c-1672193f2d16","Type":"ContainerStarted","Data":"bf41afac3ad25316b5dc2ff964d5e0f2e7fcb6553d11a4a3bd1ca65dba35dfc3"} Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.221295 4629 scope.go:117] "RemoveContainer" containerID="4a5a857a435fc978860280f4cbe148a7f4dd2ee9a3c7d2ab96099c938c4b272a" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.228841 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.236419 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.273573 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 09:01:51 crc kubenswrapper[4629]: E1211 09:01:51.273971 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c633337-5e1d-438b-8b18-0acb62ec2da0" containerName="setup-container" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.273985 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c633337-5e1d-438b-8b18-0acb62ec2da0" containerName="setup-container" Dec 11 09:01:51 crc kubenswrapper[4629]: E1211 09:01:51.274007 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c633337-5e1d-438b-8b18-0acb62ec2da0" containerName="rabbitmq" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.274013 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c633337-5e1d-438b-8b18-0acb62ec2da0" containerName="rabbitmq" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.274207 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c633337-5e1d-438b-8b18-0acb62ec2da0" containerName="rabbitmq" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.275194 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.278118 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.278404 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.279301 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.279322 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.279448 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.279509 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.279620 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qgg55" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.295217 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.368762 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82957ad9-6716-4f92-beb1-dd7502ece10d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.368805 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82957ad9-6716-4f92-beb1-dd7502ece10d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.368832 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82957ad9-6716-4f92-beb1-dd7502ece10d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.368909 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82957ad9-6716-4f92-beb1-dd7502ece10d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.368927 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqs5t\" (UniqueName: \"kubernetes.io/projected/82957ad9-6716-4f92-beb1-dd7502ece10d-kube-api-access-cqs5t\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.368941 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82957ad9-6716-4f92-beb1-dd7502ece10d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.369010 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82957ad9-6716-4f92-beb1-dd7502ece10d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.369032 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82957ad9-6716-4f92-beb1-dd7502ece10d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.369135 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82957ad9-6716-4f92-beb1-dd7502ece10d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.369180 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.369348 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82957ad9-6716-4f92-beb1-dd7502ece10d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.470480 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82957ad9-6716-4f92-beb1-dd7502ece10d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.470536 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82957ad9-6716-4f92-beb1-dd7502ece10d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.470557 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqs5t\" (UniqueName: \"kubernetes.io/projected/82957ad9-6716-4f92-beb1-dd7502ece10d-kube-api-access-cqs5t\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.470578 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82957ad9-6716-4f92-beb1-dd7502ece10d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.470644 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82957ad9-6716-4f92-beb1-dd7502ece10d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.470674 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82957ad9-6716-4f92-beb1-dd7502ece10d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.470698 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82957ad9-6716-4f92-beb1-dd7502ece10d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.471102 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82957ad9-6716-4f92-beb1-dd7502ece10d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.471160 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.471200 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82957ad9-6716-4f92-beb1-dd7502ece10d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.471467 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82957ad9-6716-4f92-beb1-dd7502ece10d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.471545 4629 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.472115 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82957ad9-6716-4f92-beb1-dd7502ece10d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.472163 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82957ad9-6716-4f92-beb1-dd7502ece10d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.472309 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82957ad9-6716-4f92-beb1-dd7502ece10d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.472622 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82957ad9-6716-4f92-beb1-dd7502ece10d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.473496 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82957ad9-6716-4f92-beb1-dd7502ece10d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.534103 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82957ad9-6716-4f92-beb1-dd7502ece10d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.534419 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82957ad9-6716-4f92-beb1-dd7502ece10d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.534754 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82957ad9-6716-4f92-beb1-dd7502ece10d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.535080 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82957ad9-6716-4f92-beb1-dd7502ece10d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.535644 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqs5t\" (UniqueName: \"kubernetes.io/projected/82957ad9-6716-4f92-beb1-dd7502ece10d-kube-api-access-cqs5t\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.659125 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"82957ad9-6716-4f92-beb1-dd7502ece10d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:51 crc kubenswrapper[4629]: I1211 09:01:51.918268 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:01:52 crc kubenswrapper[4629]: I1211 09:01:52.213343 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c633337-5e1d-438b-8b18-0acb62ec2da0" path="/var/lib/kubelet/pods/0c633337-5e1d-438b-8b18-0acb62ec2da0/volumes" Dec 11 09:01:52 crc kubenswrapper[4629]: I1211 09:01:52.214767 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8c10d7a6-61cb-4dd8-884c-1672193f2d16","Type":"ContainerStarted","Data":"50e39bd33a4ebc697dacfbf5c265b00b8ac31b227b8b77365b6d4fe8594b7c05"} Dec 11 09:01:52 crc kubenswrapper[4629]: I1211 09:01:52.394830 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 09:01:53 crc kubenswrapper[4629]: I1211 09:01:53.225964 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"82957ad9-6716-4f92-beb1-dd7502ece10d","Type":"ContainerStarted","Data":"a6f869ecfae39f0e2843872811925abcca9c0f5582091ccf83a4e26a05d68cb2"} Dec 11 09:01:54 crc kubenswrapper[4629]: I1211 09:01:54.236006 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"82957ad9-6716-4f92-beb1-dd7502ece10d","Type":"ContainerStarted","Data":"8bb60d0579a75c20b61709bf9abb957935895dcbe407b68f3ab41ab5aa77511a"} Dec 11 09:01:55 crc kubenswrapper[4629]: I1211 09:01:55.904238 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-v9hjs"] Dec 11 09:01:55 crc kubenswrapper[4629]: I1211 09:01:55.906899 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:55 crc kubenswrapper[4629]: I1211 09:01:55.909254 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 11 09:01:55 crc kubenswrapper[4629]: I1211 09:01:55.927607 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-v9hjs"] Dec 11 09:01:55 crc kubenswrapper[4629]: I1211 09:01:55.978537 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkb9w\" (UniqueName: \"kubernetes.io/projected/0888f617-5a2a-4421-8aea-5398df6ac4cd-kube-api-access-fkb9w\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:55 crc kubenswrapper[4629]: I1211 09:01:55.978584 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-config\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:55 crc kubenswrapper[4629]: I1211 09:01:55.978655 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:55 crc kubenswrapper[4629]: I1211 09:01:55.978693 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-dns-svc\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:55 crc kubenswrapper[4629]: I1211 09:01:55.978773 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:55 crc kubenswrapper[4629]: I1211 09:01:55.978812 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.079842 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.079936 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.079963 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkb9w\" (UniqueName: \"kubernetes.io/projected/0888f617-5a2a-4421-8aea-5398df6ac4cd-kube-api-access-fkb9w\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.079981 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-config\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.080041 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.080080 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-dns-svc\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.080981 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.080996 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.081129 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-config\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.081143 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-dns-svc\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.081653 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.098499 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkb9w\" (UniqueName: \"kubernetes.io/projected/0888f617-5a2a-4421-8aea-5398df6ac4cd-kube-api-access-fkb9w\") pod \"dnsmasq-dns-578b8d767c-v9hjs\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.263765 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:56 crc kubenswrapper[4629]: I1211 09:01:56.710385 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-v9hjs"] Dec 11 09:01:56 crc kubenswrapper[4629]: W1211 09:01:56.711595 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0888f617_5a2a_4421_8aea_5398df6ac4cd.slice/crio-168290c949c0291ac1fafaa67ff8385ecec3e6b5c7b8fd79c7b9b8187b544036 WatchSource:0}: Error finding container 168290c949c0291ac1fafaa67ff8385ecec3e6b5c7b8fd79c7b9b8187b544036: Status 404 returned error can't find the container with id 168290c949c0291ac1fafaa67ff8385ecec3e6b5c7b8fd79c7b9b8187b544036 Dec 11 09:01:57 crc kubenswrapper[4629]: I1211 09:01:57.260837 4629 generic.go:334] "Generic (PLEG): container finished" podID="0888f617-5a2a-4421-8aea-5398df6ac4cd" containerID="3223e622f94931029572296621c31e646daf1241c0581f6f80a8c8a2a4cd765f" exitCode=0 Dec 11 09:01:57 crc kubenswrapper[4629]: I1211 09:01:57.260991 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" event={"ID":"0888f617-5a2a-4421-8aea-5398df6ac4cd","Type":"ContainerDied","Data":"3223e622f94931029572296621c31e646daf1241c0581f6f80a8c8a2a4cd765f"} Dec 11 09:01:57 crc kubenswrapper[4629]: I1211 09:01:57.261296 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" event={"ID":"0888f617-5a2a-4421-8aea-5398df6ac4cd","Type":"ContainerStarted","Data":"168290c949c0291ac1fafaa67ff8385ecec3e6b5c7b8fd79c7b9b8187b544036"} Dec 11 09:01:58 crc kubenswrapper[4629]: I1211 09:01:58.272260 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" event={"ID":"0888f617-5a2a-4421-8aea-5398df6ac4cd","Type":"ContainerStarted","Data":"446fa68982ffd69dfb0cdbccd28c38997e4bc811cbde85b106d7c5316cb04eaf"} Dec 11 09:01:58 crc kubenswrapper[4629]: I1211 09:01:58.273941 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:01:58 crc kubenswrapper[4629]: I1211 09:01:58.292769 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" podStartSLOduration=3.2927537 podStartE2EDuration="3.2927537s" podCreationTimestamp="2025-12-11 09:01:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:01:58.289904481 +0000 UTC m=+1506.355320089" watchObservedRunningTime="2025-12-11 09:01:58.2927537 +0000 UTC m=+1506.358169308" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.265272 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.338248 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-zkqrp"] Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.338814 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" podUID="d17744e4-f52a-4c2a-8a59-628d6e506f49" containerName="dnsmasq-dns" containerID="cri-o://c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915" gracePeriod=10 Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.575804 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-667ff9c869-m6rlw"] Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.577384 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.601119 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-667ff9c869-m6rlw"] Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.656021 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" podUID="d17744e4-f52a-4c2a-8a59-628d6e506f49" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.180:5353: connect: connection refused" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.693288 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spb2b\" (UniqueName: \"kubernetes.io/projected/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-kube-api-access-spb2b\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.693374 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-config\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.693433 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-openstack-edpm-ipam\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.693517 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-ovsdbserver-sb\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.693592 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-ovsdbserver-nb\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.693917 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-dns-svc\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.796799 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-dns-svc\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.796915 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spb2b\" (UniqueName: \"kubernetes.io/projected/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-kube-api-access-spb2b\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.796947 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-config\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.796988 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-openstack-edpm-ipam\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.797045 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-ovsdbserver-sb\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.797066 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-ovsdbserver-nb\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.798060 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-ovsdbserver-nb\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.798115 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-config\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.798536 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-openstack-edpm-ipam\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.798914 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-dns-svc\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.800190 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-ovsdbserver-sb\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.830212 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spb2b\" (UniqueName: \"kubernetes.io/projected/12aa6e1b-5b39-48c9-80bf-6f57d8f5333e-kube-api-access-spb2b\") pod \"dnsmasq-dns-667ff9c869-m6rlw\" (UID: \"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e\") " pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:06 crc kubenswrapper[4629]: I1211 09:02:06.910671 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:06.994518 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.103593 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-dns-svc\") pod \"d17744e4-f52a-4c2a-8a59-628d6e506f49\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.104061 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-ovsdbserver-sb\") pod \"d17744e4-f52a-4c2a-8a59-628d6e506f49\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.104116 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-config\") pod \"d17744e4-f52a-4c2a-8a59-628d6e506f49\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.104392 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-ovsdbserver-nb\") pod \"d17744e4-f52a-4c2a-8a59-628d6e506f49\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.104457 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrztm\" (UniqueName: \"kubernetes.io/projected/d17744e4-f52a-4c2a-8a59-628d6e506f49-kube-api-access-jrztm\") pod \"d17744e4-f52a-4c2a-8a59-628d6e506f49\" (UID: \"d17744e4-f52a-4c2a-8a59-628d6e506f49\") " Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.150163 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d17744e4-f52a-4c2a-8a59-628d6e506f49-kube-api-access-jrztm" (OuterVolumeSpecName: "kube-api-access-jrztm") pod "d17744e4-f52a-4c2a-8a59-628d6e506f49" (UID: "d17744e4-f52a-4c2a-8a59-628d6e506f49"). InnerVolumeSpecName "kube-api-access-jrztm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.207277 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-config" (OuterVolumeSpecName: "config") pod "d17744e4-f52a-4c2a-8a59-628d6e506f49" (UID: "d17744e4-f52a-4c2a-8a59-628d6e506f49"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.210648 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-config\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.210687 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrztm\" (UniqueName: \"kubernetes.io/projected/d17744e4-f52a-4c2a-8a59-628d6e506f49-kube-api-access-jrztm\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.263658 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d17744e4-f52a-4c2a-8a59-628d6e506f49" (UID: "d17744e4-f52a-4c2a-8a59-628d6e506f49"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.291545 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d17744e4-f52a-4c2a-8a59-628d6e506f49" (UID: "d17744e4-f52a-4c2a-8a59-628d6e506f49"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.317246 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.317285 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.317442 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d17744e4-f52a-4c2a-8a59-628d6e506f49" (UID: "d17744e4-f52a-4c2a-8a59-628d6e506f49"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.355873 4629 generic.go:334] "Generic (PLEG): container finished" podID="d17744e4-f52a-4c2a-8a59-628d6e506f49" containerID="c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915" exitCode=0 Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.355932 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" event={"ID":"d17744e4-f52a-4c2a-8a59-628d6e506f49","Type":"ContainerDied","Data":"c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915"} Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.355967 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.355972 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-zkqrp" event={"ID":"d17744e4-f52a-4c2a-8a59-628d6e506f49","Type":"ContainerDied","Data":"ddac5ffded1d4e1db29805322af71a6e528cd7264d9bf447fc122f4226ce67e4"} Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.355982 4629 scope.go:117] "RemoveContainer" containerID="c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.422096 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d17744e4-f52a-4c2a-8a59-628d6e506f49-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.424058 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-zkqrp"] Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.451010 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-zkqrp"] Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.463097 4629 scope.go:117] "RemoveContainer" containerID="ddd9af99d2ea9a8c5027dd523d564f83dc8b382932cdebccae7a4d8d8aa9933c" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.582475 4629 scope.go:117] "RemoveContainer" containerID="c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915" Dec 11 09:02:07 crc kubenswrapper[4629]: E1211 09:02:07.582918 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915\": container with ID starting with c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915 not found: ID does not exist" containerID="c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.582954 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915"} err="failed to get container status \"c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915\": rpc error: code = NotFound desc = could not find container \"c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915\": container with ID starting with c6ff1636aa5b84d09cb2000908d9d39b45dad164d57e7982e350ad6557deb915 not found: ID does not exist" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.582978 4629 scope.go:117] "RemoveContainer" containerID="ddd9af99d2ea9a8c5027dd523d564f83dc8b382932cdebccae7a4d8d8aa9933c" Dec 11 09:02:07 crc kubenswrapper[4629]: E1211 09:02:07.583345 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddd9af99d2ea9a8c5027dd523d564f83dc8b382932cdebccae7a4d8d8aa9933c\": container with ID starting with ddd9af99d2ea9a8c5027dd523d564f83dc8b382932cdebccae7a4d8d8aa9933c not found: ID does not exist" containerID="ddd9af99d2ea9a8c5027dd523d564f83dc8b382932cdebccae7a4d8d8aa9933c" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.583372 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddd9af99d2ea9a8c5027dd523d564f83dc8b382932cdebccae7a4d8d8aa9933c"} err="failed to get container status \"ddd9af99d2ea9a8c5027dd523d564f83dc8b382932cdebccae7a4d8d8aa9933c\": rpc error: code = NotFound desc = could not find container \"ddd9af99d2ea9a8c5027dd523d564f83dc8b382932cdebccae7a4d8d8aa9933c\": container with ID starting with ddd9af99d2ea9a8c5027dd523d564f83dc8b382932cdebccae7a4d8d8aa9933c not found: ID does not exist" Dec 11 09:02:07 crc kubenswrapper[4629]: I1211 09:02:07.651109 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-667ff9c869-m6rlw"] Dec 11 09:02:08 crc kubenswrapper[4629]: E1211 09:02:08.142323 4629 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12aa6e1b_5b39_48c9_80bf_6f57d8f5333e.slice/crio-conmon-6a4114923b507de60cdd0a8f72c7d279df3bb22848c67495a762cbdf086bfb13.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12aa6e1b_5b39_48c9_80bf_6f57d8f5333e.slice/crio-6a4114923b507de60cdd0a8f72c7d279df3bb22848c67495a762cbdf086bfb13.scope\": RecentStats: unable to find data in memory cache]" Dec 11 09:02:08 crc kubenswrapper[4629]: I1211 09:02:08.209546 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d17744e4-f52a-4c2a-8a59-628d6e506f49" path="/var/lib/kubelet/pods/d17744e4-f52a-4c2a-8a59-628d6e506f49/volumes" Dec 11 09:02:08 crc kubenswrapper[4629]: I1211 09:02:08.365807 4629 generic.go:334] "Generic (PLEG): container finished" podID="12aa6e1b-5b39-48c9-80bf-6f57d8f5333e" containerID="6a4114923b507de60cdd0a8f72c7d279df3bb22848c67495a762cbdf086bfb13" exitCode=0 Dec 11 09:02:08 crc kubenswrapper[4629]: I1211 09:02:08.365886 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" event={"ID":"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e","Type":"ContainerDied","Data":"6a4114923b507de60cdd0a8f72c7d279df3bb22848c67495a762cbdf086bfb13"} Dec 11 09:02:08 crc kubenswrapper[4629]: I1211 09:02:08.365913 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" event={"ID":"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e","Type":"ContainerStarted","Data":"26acde0b3f40259b67cd2524fcd2dd900cffd795f2af8925a7e12c168938813d"} Dec 11 09:02:09 crc kubenswrapper[4629]: I1211 09:02:09.381395 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" event={"ID":"12aa6e1b-5b39-48c9-80bf-6f57d8f5333e","Type":"ContainerStarted","Data":"521de67c5f25ea8b60603d3b8c2c65b316f8c6926e8b0f3690101c91c9c75814"} Dec 11 09:02:09 crc kubenswrapper[4629]: I1211 09:02:09.381888 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:09 crc kubenswrapper[4629]: I1211 09:02:09.418261 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" podStartSLOduration=3.418241657 podStartE2EDuration="3.418241657s" podCreationTimestamp="2025-12-11 09:02:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:02:09.408494271 +0000 UTC m=+1517.473909879" watchObservedRunningTime="2025-12-11 09:02:09.418241657 +0000 UTC m=+1517.483657265" Dec 11 09:02:16 crc kubenswrapper[4629]: I1211 09:02:16.913120 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-667ff9c869-m6rlw" Dec 11 09:02:16 crc kubenswrapper[4629]: I1211 09:02:16.973645 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-v9hjs"] Dec 11 09:02:16 crc kubenswrapper[4629]: I1211 09:02:16.974006 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" podUID="0888f617-5a2a-4421-8aea-5398df6ac4cd" containerName="dnsmasq-dns" containerID="cri-o://446fa68982ffd69dfb0cdbccd28c38997e4bc811cbde85b106d7c5316cb04eaf" gracePeriod=10 Dec 11 09:02:17 crc kubenswrapper[4629]: I1211 09:02:17.457730 4629 generic.go:334] "Generic (PLEG): container finished" podID="0888f617-5a2a-4421-8aea-5398df6ac4cd" containerID="446fa68982ffd69dfb0cdbccd28c38997e4bc811cbde85b106d7c5316cb04eaf" exitCode=0 Dec 11 09:02:17 crc kubenswrapper[4629]: I1211 09:02:17.457788 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" event={"ID":"0888f617-5a2a-4421-8aea-5398df6ac4cd","Type":"ContainerDied","Data":"446fa68982ffd69dfb0cdbccd28c38997e4bc811cbde85b106d7c5316cb04eaf"} Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.222156 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.357788 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkb9w\" (UniqueName: \"kubernetes.io/projected/0888f617-5a2a-4421-8aea-5398df6ac4cd-kube-api-access-fkb9w\") pod \"0888f617-5a2a-4421-8aea-5398df6ac4cd\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.359222 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-config\") pod \"0888f617-5a2a-4421-8aea-5398df6ac4cd\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.359314 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-dns-svc\") pod \"0888f617-5a2a-4421-8aea-5398df6ac4cd\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.359366 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-ovsdbserver-nb\") pod \"0888f617-5a2a-4421-8aea-5398df6ac4cd\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.359441 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-ovsdbserver-sb\") pod \"0888f617-5a2a-4421-8aea-5398df6ac4cd\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.359478 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-openstack-edpm-ipam\") pod \"0888f617-5a2a-4421-8aea-5398df6ac4cd\" (UID: \"0888f617-5a2a-4421-8aea-5398df6ac4cd\") " Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.592236 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" event={"ID":"0888f617-5a2a-4421-8aea-5398df6ac4cd","Type":"ContainerDied","Data":"168290c949c0291ac1fafaa67ff8385ecec3e6b5c7b8fd79c7b9b8187b544036"} Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.592589 4629 scope.go:117] "RemoveContainer" containerID="446fa68982ffd69dfb0cdbccd28c38997e4bc811cbde85b106d7c5316cb04eaf" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.592808 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-v9hjs" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.606108 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0888f617-5a2a-4421-8aea-5398df6ac4cd-kube-api-access-fkb9w" (OuterVolumeSpecName: "kube-api-access-fkb9w") pod "0888f617-5a2a-4421-8aea-5398df6ac4cd" (UID: "0888f617-5a2a-4421-8aea-5398df6ac4cd"). InnerVolumeSpecName "kube-api-access-fkb9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.642188 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkb9w\" (UniqueName: \"kubernetes.io/projected/0888f617-5a2a-4421-8aea-5398df6ac4cd-kube-api-access-fkb9w\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.657275 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "0888f617-5a2a-4421-8aea-5398df6ac4cd" (UID: "0888f617-5a2a-4421-8aea-5398df6ac4cd"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.687387 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0888f617-5a2a-4421-8aea-5398df6ac4cd" (UID: "0888f617-5a2a-4421-8aea-5398df6ac4cd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.707157 4629 scope.go:117] "RemoveContainer" containerID="3223e622f94931029572296621c31e646daf1241c0581f6f80a8c8a2a4cd765f" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.709758 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0888f617-5a2a-4421-8aea-5398df6ac4cd" (UID: "0888f617-5a2a-4421-8aea-5398df6ac4cd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.717259 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-config" (OuterVolumeSpecName: "config") pod "0888f617-5a2a-4421-8aea-5398df6ac4cd" (UID: "0888f617-5a2a-4421-8aea-5398df6ac4cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.718972 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0888f617-5a2a-4421-8aea-5398df6ac4cd" (UID: "0888f617-5a2a-4421-8aea-5398df6ac4cd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.747442 4629 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.747483 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.747497 4629 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.747511 4629 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.747534 4629 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0888f617-5a2a-4421-8aea-5398df6ac4cd-config\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.929361 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-v9hjs"] Dec 11 09:02:18 crc kubenswrapper[4629]: I1211 09:02:18.948112 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-v9hjs"] Dec 11 09:02:20 crc kubenswrapper[4629]: I1211 09:02:20.209980 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0888f617-5a2a-4421-8aea-5398df6ac4cd" path="/var/lib/kubelet/pods/0888f617-5a2a-4421-8aea-5398df6ac4cd/volumes" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.729821 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lj8zh"] Dec 11 09:02:21 crc kubenswrapper[4629]: E1211 09:02:21.730646 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0888f617-5a2a-4421-8aea-5398df6ac4cd" containerName="dnsmasq-dns" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.730664 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="0888f617-5a2a-4421-8aea-5398df6ac4cd" containerName="dnsmasq-dns" Dec 11 09:02:21 crc kubenswrapper[4629]: E1211 09:02:21.730689 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17744e4-f52a-4c2a-8a59-628d6e506f49" containerName="init" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.730696 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17744e4-f52a-4c2a-8a59-628d6e506f49" containerName="init" Dec 11 09:02:21 crc kubenswrapper[4629]: E1211 09:02:21.730718 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0888f617-5a2a-4421-8aea-5398df6ac4cd" containerName="init" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.730728 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="0888f617-5a2a-4421-8aea-5398df6ac4cd" containerName="init" Dec 11 09:02:21 crc kubenswrapper[4629]: E1211 09:02:21.730752 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17744e4-f52a-4c2a-8a59-628d6e506f49" containerName="dnsmasq-dns" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.730762 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17744e4-f52a-4c2a-8a59-628d6e506f49" containerName="dnsmasq-dns" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.730993 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="0888f617-5a2a-4421-8aea-5398df6ac4cd" containerName="dnsmasq-dns" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.731018 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17744e4-f52a-4c2a-8a59-628d6e506f49" containerName="dnsmasq-dns" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.732613 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.749834 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj8zh"] Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.875170 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3c941f0-8454-4f45-9bf9-cf2089612200-catalog-content\") pod \"redhat-marketplace-lj8zh\" (UID: \"f3c941f0-8454-4f45-9bf9-cf2089612200\") " pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.875260 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgghj\" (UniqueName: \"kubernetes.io/projected/f3c941f0-8454-4f45-9bf9-cf2089612200-kube-api-access-lgghj\") pod \"redhat-marketplace-lj8zh\" (UID: \"f3c941f0-8454-4f45-9bf9-cf2089612200\") " pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.875319 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3c941f0-8454-4f45-9bf9-cf2089612200-utilities\") pod \"redhat-marketplace-lj8zh\" (UID: \"f3c941f0-8454-4f45-9bf9-cf2089612200\") " pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.977655 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3c941f0-8454-4f45-9bf9-cf2089612200-catalog-content\") pod \"redhat-marketplace-lj8zh\" (UID: \"f3c941f0-8454-4f45-9bf9-cf2089612200\") " pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.977755 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgghj\" (UniqueName: \"kubernetes.io/projected/f3c941f0-8454-4f45-9bf9-cf2089612200-kube-api-access-lgghj\") pod \"redhat-marketplace-lj8zh\" (UID: \"f3c941f0-8454-4f45-9bf9-cf2089612200\") " pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.977816 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3c941f0-8454-4f45-9bf9-cf2089612200-utilities\") pod \"redhat-marketplace-lj8zh\" (UID: \"f3c941f0-8454-4f45-9bf9-cf2089612200\") " pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.978340 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3c941f0-8454-4f45-9bf9-cf2089612200-catalog-content\") pod \"redhat-marketplace-lj8zh\" (UID: \"f3c941f0-8454-4f45-9bf9-cf2089612200\") " pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.978384 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3c941f0-8454-4f45-9bf9-cf2089612200-utilities\") pod \"redhat-marketplace-lj8zh\" (UID: \"f3c941f0-8454-4f45-9bf9-cf2089612200\") " pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:21 crc kubenswrapper[4629]: I1211 09:02:21.997984 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgghj\" (UniqueName: \"kubernetes.io/projected/f3c941f0-8454-4f45-9bf9-cf2089612200-kube-api-access-lgghj\") pod \"redhat-marketplace-lj8zh\" (UID: \"f3c941f0-8454-4f45-9bf9-cf2089612200\") " pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.100184 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.712989 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf"] Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.714417 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.716767 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.717044 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.717331 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.719449 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qjfc4" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.751039 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf"] Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.798653 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj8zh"] Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.811176 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.811422 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.811475 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.811525 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgsgh\" (UniqueName: \"kubernetes.io/projected/0206ce20-df99-494c-96f0-f6201de98376-kube-api-access-sgsgh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.914179 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.914752 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.914790 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.914824 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgsgh\" (UniqueName: \"kubernetes.io/projected/0206ce20-df99-494c-96f0-f6201de98376-kube-api-access-sgsgh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.925106 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.925840 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.930207 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:22 crc kubenswrapper[4629]: I1211 09:02:22.943571 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgsgh\" (UniqueName: \"kubernetes.io/projected/0206ce20-df99-494c-96f0-f6201de98376-kube-api-access-sgsgh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:23 crc kubenswrapper[4629]: I1211 09:02:23.034824 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:02:23 crc kubenswrapper[4629]: I1211 09:02:23.639211 4629 generic.go:334] "Generic (PLEG): container finished" podID="f3c941f0-8454-4f45-9bf9-cf2089612200" containerID="f5a9311d22ff6637bda201980c98c5f37d52db2e1ec3ba22a49c444fad2e8990" exitCode=0 Dec 11 09:02:23 crc kubenswrapper[4629]: I1211 09:02:23.639756 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8zh" event={"ID":"f3c941f0-8454-4f45-9bf9-cf2089612200","Type":"ContainerDied","Data":"f5a9311d22ff6637bda201980c98c5f37d52db2e1ec3ba22a49c444fad2e8990"} Dec 11 09:02:23 crc kubenswrapper[4629]: I1211 09:02:23.639790 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8zh" event={"ID":"f3c941f0-8454-4f45-9bf9-cf2089612200","Type":"ContainerStarted","Data":"f7e7c8182d1eb5062f0801f4f7f9cb388a23c1b93aa67c4ed1b61d9c2db6f930"} Dec 11 09:02:23 crc kubenswrapper[4629]: I1211 09:02:23.903973 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf"] Dec 11 09:02:23 crc kubenswrapper[4629]: W1211 09:02:23.912626 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0206ce20_df99_494c_96f0_f6201de98376.slice/crio-85278d128cb805fbd31666e91c64e269fe8dd6413ed5bb8e6b8a6e11a7a4744b WatchSource:0}: Error finding container 85278d128cb805fbd31666e91c64e269fe8dd6413ed5bb8e6b8a6e11a7a4744b: Status 404 returned error can't find the container with id 85278d128cb805fbd31666e91c64e269fe8dd6413ed5bb8e6b8a6e11a7a4744b Dec 11 09:02:24 crc kubenswrapper[4629]: I1211 09:02:24.648087 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" event={"ID":"0206ce20-df99-494c-96f0-f6201de98376","Type":"ContainerStarted","Data":"85278d128cb805fbd31666e91c64e269fe8dd6413ed5bb8e6b8a6e11a7a4744b"} Dec 11 09:02:24 crc kubenswrapper[4629]: I1211 09:02:24.651012 4629 generic.go:334] "Generic (PLEG): container finished" podID="8c10d7a6-61cb-4dd8-884c-1672193f2d16" containerID="50e39bd33a4ebc697dacfbf5c265b00b8ac31b227b8b77365b6d4fe8594b7c05" exitCode=0 Dec 11 09:02:24 crc kubenswrapper[4629]: I1211 09:02:24.651041 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8c10d7a6-61cb-4dd8-884c-1672193f2d16","Type":"ContainerDied","Data":"50e39bd33a4ebc697dacfbf5c265b00b8ac31b227b8b77365b6d4fe8594b7c05"} Dec 11 09:02:25 crc kubenswrapper[4629]: I1211 09:02:25.679424 4629 generic.go:334] "Generic (PLEG): container finished" podID="f3c941f0-8454-4f45-9bf9-cf2089612200" containerID="abf3eada544f84a55475144fad82f7a00c9a176d92569c1f008550f1b93261d2" exitCode=0 Dec 11 09:02:25 crc kubenswrapper[4629]: I1211 09:02:25.679923 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8zh" event={"ID":"f3c941f0-8454-4f45-9bf9-cf2089612200","Type":"ContainerDied","Data":"abf3eada544f84a55475144fad82f7a00c9a176d92569c1f008550f1b93261d2"} Dec 11 09:02:25 crc kubenswrapper[4629]: I1211 09:02:25.688429 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8c10d7a6-61cb-4dd8-884c-1672193f2d16","Type":"ContainerStarted","Data":"65e7e63cfeaf5a3bcd0de7ee6fc6eb21329e25b974bb65627d3b9618127084f4"} Dec 11 09:02:25 crc kubenswrapper[4629]: I1211 09:02:25.689168 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 11 09:02:25 crc kubenswrapper[4629]: I1211 09:02:25.751184 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.75112396 podStartE2EDuration="36.75112396s" podCreationTimestamp="2025-12-11 09:01:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:02:25.74224307 +0000 UTC m=+1533.807658678" watchObservedRunningTime="2025-12-11 09:02:25.75112396 +0000 UTC m=+1533.816539568" Dec 11 09:02:26 crc kubenswrapper[4629]: I1211 09:02:26.698640 4629 generic.go:334] "Generic (PLEG): container finished" podID="82957ad9-6716-4f92-beb1-dd7502ece10d" containerID="8bb60d0579a75c20b61709bf9abb957935895dcbe407b68f3ab41ab5aa77511a" exitCode=0 Dec 11 09:02:26 crc kubenswrapper[4629]: I1211 09:02:26.698723 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"82957ad9-6716-4f92-beb1-dd7502ece10d","Type":"ContainerDied","Data":"8bb60d0579a75c20b61709bf9abb957935895dcbe407b68f3ab41ab5aa77511a"} Dec 11 09:02:26 crc kubenswrapper[4629]: I1211 09:02:26.702938 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8zh" event={"ID":"f3c941f0-8454-4f45-9bf9-cf2089612200","Type":"ContainerStarted","Data":"2bf897b1911ef6e398229db095544f30b5cd6e7d7e2a091bdec4200a56f89b1a"} Dec 11 09:02:26 crc kubenswrapper[4629]: I1211 09:02:26.809013 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lj8zh" podStartSLOduration=3.258745745 podStartE2EDuration="5.808988699s" podCreationTimestamp="2025-12-11 09:02:21 +0000 UTC" firstStartedPulling="2025-12-11 09:02:23.642892026 +0000 UTC m=+1531.708307644" lastFinishedPulling="2025-12-11 09:02:26.19313499 +0000 UTC m=+1534.258550598" observedRunningTime="2025-12-11 09:02:26.796690912 +0000 UTC m=+1534.862106530" watchObservedRunningTime="2025-12-11 09:02:26.808988699 +0000 UTC m=+1534.874404307" Dec 11 09:02:27 crc kubenswrapper[4629]: I1211 09:02:27.725931 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"82957ad9-6716-4f92-beb1-dd7502ece10d","Type":"ContainerStarted","Data":"13ffe8c691fbc46d579573ecae2eeb0c7e8bfae930ff008cc7b31a5cfb704011"} Dec 11 09:02:27 crc kubenswrapper[4629]: I1211 09:02:27.726878 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:02:27 crc kubenswrapper[4629]: I1211 09:02:27.770837 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.770820872 podStartE2EDuration="36.770820872s" podCreationTimestamp="2025-12-11 09:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:02:27.76882661 +0000 UTC m=+1535.834242228" watchObservedRunningTime="2025-12-11 09:02:27.770820872 +0000 UTC m=+1535.836236480" Dec 11 09:02:32 crc kubenswrapper[4629]: I1211 09:02:32.190051 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:32 crc kubenswrapper[4629]: I1211 09:02:32.190554 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:32 crc kubenswrapper[4629]: I1211 09:02:32.312078 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:32 crc kubenswrapper[4629]: I1211 09:02:32.856279 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:32 crc kubenswrapper[4629]: I1211 09:02:32.947585 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj8zh"] Dec 11 09:02:34 crc kubenswrapper[4629]: I1211 09:02:34.826151 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lj8zh" podUID="f3c941f0-8454-4f45-9bf9-cf2089612200" containerName="registry-server" containerID="cri-o://2bf897b1911ef6e398229db095544f30b5cd6e7d7e2a091bdec4200a56f89b1a" gracePeriod=2 Dec 11 09:02:35 crc kubenswrapper[4629]: I1211 09:02:35.851049 4629 generic.go:334] "Generic (PLEG): container finished" podID="f3c941f0-8454-4f45-9bf9-cf2089612200" containerID="2bf897b1911ef6e398229db095544f30b5cd6e7d7e2a091bdec4200a56f89b1a" exitCode=0 Dec 11 09:02:35 crc kubenswrapper[4629]: I1211 09:02:35.851078 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8zh" event={"ID":"f3c941f0-8454-4f45-9bf9-cf2089612200","Type":"ContainerDied","Data":"2bf897b1911ef6e398229db095544f30b5cd6e7d7e2a091bdec4200a56f89b1a"} Dec 11 09:02:39 crc kubenswrapper[4629]: I1211 09:02:39.868725 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="8c10d7a6-61cb-4dd8-884c-1672193f2d16" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.189:5671: connect: connection refused" Dec 11 09:02:41 crc kubenswrapper[4629]: I1211 09:02:41.921423 4629 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="82957ad9-6716-4f92-beb1-dd7502ece10d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.190:5671: connect: connection refused" Dec 11 09:02:42 crc kubenswrapper[4629]: E1211 09:02:42.191073 4629 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2bf897b1911ef6e398229db095544f30b5cd6e7d7e2a091bdec4200a56f89b1a is running failed: container process not found" containerID="2bf897b1911ef6e398229db095544f30b5cd6e7d7e2a091bdec4200a56f89b1a" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 09:02:42 crc kubenswrapper[4629]: E1211 09:02:42.191428 4629 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2bf897b1911ef6e398229db095544f30b5cd6e7d7e2a091bdec4200a56f89b1a is running failed: container process not found" containerID="2bf897b1911ef6e398229db095544f30b5cd6e7d7e2a091bdec4200a56f89b1a" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 09:02:42 crc kubenswrapper[4629]: E1211 09:02:42.191676 4629 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2bf897b1911ef6e398229db095544f30b5cd6e7d7e2a091bdec4200a56f89b1a is running failed: container process not found" containerID="2bf897b1911ef6e398229db095544f30b5cd6e7d7e2a091bdec4200a56f89b1a" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 09:02:42 crc kubenswrapper[4629]: E1211 09:02:42.191717 4629 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2bf897b1911ef6e398229db095544f30b5cd6e7d7e2a091bdec4200a56f89b1a is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-lj8zh" podUID="f3c941f0-8454-4f45-9bf9-cf2089612200" containerName="registry-server" Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.523148 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:45 crc kubenswrapper[4629]: E1211 09:02:45.529988 4629 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Dec 11 09:02:45 crc kubenswrapper[4629]: E1211 09:02:45.530151 4629 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 11 09:02:45 crc kubenswrapper[4629]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Dec 11 09:02:45 crc kubenswrapper[4629]: - hosts: all Dec 11 09:02:45 crc kubenswrapper[4629]: strategy: linear Dec 11 09:02:45 crc kubenswrapper[4629]: tasks: Dec 11 09:02:45 crc kubenswrapper[4629]: - name: Enable podified-repos Dec 11 09:02:45 crc kubenswrapper[4629]: become: true Dec 11 09:02:45 crc kubenswrapper[4629]: ansible.builtin.shell: | Dec 11 09:02:45 crc kubenswrapper[4629]: set -euxo pipefail Dec 11 09:02:45 crc kubenswrapper[4629]: pushd /var/tmp Dec 11 09:02:45 crc kubenswrapper[4629]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Dec 11 09:02:45 crc kubenswrapper[4629]: pushd repo-setup-main Dec 11 09:02:45 crc kubenswrapper[4629]: python3 -m venv ./venv Dec 11 09:02:45 crc kubenswrapper[4629]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Dec 11 09:02:45 crc kubenswrapper[4629]: ./venv/bin/repo-setup current-podified -b antelope Dec 11 09:02:45 crc kubenswrapper[4629]: popd Dec 11 09:02:45 crc kubenswrapper[4629]: rm -rf repo-setup-main Dec 11 09:02:45 crc kubenswrapper[4629]: Dec 11 09:02:45 crc kubenswrapper[4629]: Dec 11 09:02:45 crc kubenswrapper[4629]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Dec 11 09:02:45 crc kubenswrapper[4629]: edpm_override_hosts: openstack-edpm-ipam Dec 11 09:02:45 crc kubenswrapper[4629]: edpm_service_type: repo-setup Dec 11 09:02:45 crc kubenswrapper[4629]: Dec 11 09:02:45 crc kubenswrapper[4629]: Dec 11 09:02:45 crc kubenswrapper[4629]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sgsgh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf_openstack(0206ce20-df99-494c-96f0-f6201de98376): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Dec 11 09:02:45 crc kubenswrapper[4629]: > logger="UnhandledError" Dec 11 09:02:45 crc kubenswrapper[4629]: E1211 09:02:45.531371 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" podUID="0206ce20-df99-494c-96f0-f6201de98376" Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.675069 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3c941f0-8454-4f45-9bf9-cf2089612200-utilities\") pod \"f3c941f0-8454-4f45-9bf9-cf2089612200\" (UID: \"f3c941f0-8454-4f45-9bf9-cf2089612200\") " Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.675119 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3c941f0-8454-4f45-9bf9-cf2089612200-catalog-content\") pod \"f3c941f0-8454-4f45-9bf9-cf2089612200\" (UID: \"f3c941f0-8454-4f45-9bf9-cf2089612200\") " Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.675303 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgghj\" (UniqueName: \"kubernetes.io/projected/f3c941f0-8454-4f45-9bf9-cf2089612200-kube-api-access-lgghj\") pod \"f3c941f0-8454-4f45-9bf9-cf2089612200\" (UID: \"f3c941f0-8454-4f45-9bf9-cf2089612200\") " Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.675956 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3c941f0-8454-4f45-9bf9-cf2089612200-utilities" (OuterVolumeSpecName: "utilities") pod "f3c941f0-8454-4f45-9bf9-cf2089612200" (UID: "f3c941f0-8454-4f45-9bf9-cf2089612200"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.690162 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3c941f0-8454-4f45-9bf9-cf2089612200-kube-api-access-lgghj" (OuterVolumeSpecName: "kube-api-access-lgghj") pod "f3c941f0-8454-4f45-9bf9-cf2089612200" (UID: "f3c941f0-8454-4f45-9bf9-cf2089612200"). InnerVolumeSpecName "kube-api-access-lgghj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.699820 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3c941f0-8454-4f45-9bf9-cf2089612200-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3c941f0-8454-4f45-9bf9-cf2089612200" (UID: "f3c941f0-8454-4f45-9bf9-cf2089612200"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.777861 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgghj\" (UniqueName: \"kubernetes.io/projected/f3c941f0-8454-4f45-9bf9-cf2089612200-kube-api-access-lgghj\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.777906 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3c941f0-8454-4f45-9bf9-cf2089612200-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.777916 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3c941f0-8454-4f45-9bf9-cf2089612200-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.942083 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8zh" event={"ID":"f3c941f0-8454-4f45-9bf9-cf2089612200","Type":"ContainerDied","Data":"f7e7c8182d1eb5062f0801f4f7f9cb388a23c1b93aa67c4ed1b61d9c2db6f930"} Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.942135 4629 scope.go:117] "RemoveContainer" containerID="2bf897b1911ef6e398229db095544f30b5cd6e7d7e2a091bdec4200a56f89b1a" Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.942070 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lj8zh" Dec 11 09:02:45 crc kubenswrapper[4629]: E1211 09:02:45.942963 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" podUID="0206ce20-df99-494c-96f0-f6201de98376" Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.989833 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj8zh"] Dec 11 09:02:45 crc kubenswrapper[4629]: I1211 09:02:45.999199 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj8zh"] Dec 11 09:02:46 crc kubenswrapper[4629]: I1211 09:02:46.170558 4629 scope.go:117] "RemoveContainer" containerID="abf3eada544f84a55475144fad82f7a00c9a176d92569c1f008550f1b93261d2" Dec 11 09:02:46 crc kubenswrapper[4629]: I1211 09:02:46.212874 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3c941f0-8454-4f45-9bf9-cf2089612200" path="/var/lib/kubelet/pods/f3c941f0-8454-4f45-9bf9-cf2089612200/volumes" Dec 11 09:02:46 crc kubenswrapper[4629]: I1211 09:02:46.279693 4629 scope.go:117] "RemoveContainer" containerID="f5a9311d22ff6637bda201980c98c5f37d52db2e1ec3ba22a49c444fad2e8990" Dec 11 09:02:47 crc kubenswrapper[4629]: I1211 09:02:47.788342 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kf76h"] Dec 11 09:02:47 crc kubenswrapper[4629]: E1211 09:02:47.789090 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3c941f0-8454-4f45-9bf9-cf2089612200" containerName="extract-utilities" Dec 11 09:02:47 crc kubenswrapper[4629]: I1211 09:02:47.789109 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3c941f0-8454-4f45-9bf9-cf2089612200" containerName="extract-utilities" Dec 11 09:02:47 crc kubenswrapper[4629]: E1211 09:02:47.789133 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3c941f0-8454-4f45-9bf9-cf2089612200" containerName="registry-server" Dec 11 09:02:47 crc kubenswrapper[4629]: I1211 09:02:47.789139 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3c941f0-8454-4f45-9bf9-cf2089612200" containerName="registry-server" Dec 11 09:02:47 crc kubenswrapper[4629]: E1211 09:02:47.789154 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3c941f0-8454-4f45-9bf9-cf2089612200" containerName="extract-content" Dec 11 09:02:47 crc kubenswrapper[4629]: I1211 09:02:47.789160 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3c941f0-8454-4f45-9bf9-cf2089612200" containerName="extract-content" Dec 11 09:02:47 crc kubenswrapper[4629]: I1211 09:02:47.789335 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3c941f0-8454-4f45-9bf9-cf2089612200" containerName="registry-server" Dec 11 09:02:47 crc kubenswrapper[4629]: I1211 09:02:47.790692 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:02:47 crc kubenswrapper[4629]: I1211 09:02:47.817008 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kf76h"] Dec 11 09:02:47 crc kubenswrapper[4629]: I1211 09:02:47.919105 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22d09ce1-4677-427b-8007-f8743ee036dc-utilities\") pod \"certified-operators-kf76h\" (UID: \"22d09ce1-4677-427b-8007-f8743ee036dc\") " pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:02:47 crc kubenswrapper[4629]: I1211 09:02:47.919544 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22d09ce1-4677-427b-8007-f8743ee036dc-catalog-content\") pod \"certified-operators-kf76h\" (UID: \"22d09ce1-4677-427b-8007-f8743ee036dc\") " pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:02:47 crc kubenswrapper[4629]: I1211 09:02:47.919587 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm5cc\" (UniqueName: \"kubernetes.io/projected/22d09ce1-4677-427b-8007-f8743ee036dc-kube-api-access-nm5cc\") pod \"certified-operators-kf76h\" (UID: \"22d09ce1-4677-427b-8007-f8743ee036dc\") " pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:02:48 crc kubenswrapper[4629]: I1211 09:02:48.021808 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22d09ce1-4677-427b-8007-f8743ee036dc-catalog-content\") pod \"certified-operators-kf76h\" (UID: \"22d09ce1-4677-427b-8007-f8743ee036dc\") " pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:02:48 crc kubenswrapper[4629]: I1211 09:02:48.021912 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm5cc\" (UniqueName: \"kubernetes.io/projected/22d09ce1-4677-427b-8007-f8743ee036dc-kube-api-access-nm5cc\") pod \"certified-operators-kf76h\" (UID: \"22d09ce1-4677-427b-8007-f8743ee036dc\") " pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:02:48 crc kubenswrapper[4629]: I1211 09:02:48.021960 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22d09ce1-4677-427b-8007-f8743ee036dc-utilities\") pod \"certified-operators-kf76h\" (UID: \"22d09ce1-4677-427b-8007-f8743ee036dc\") " pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:02:48 crc kubenswrapper[4629]: I1211 09:02:48.022640 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22d09ce1-4677-427b-8007-f8743ee036dc-utilities\") pod \"certified-operators-kf76h\" (UID: \"22d09ce1-4677-427b-8007-f8743ee036dc\") " pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:02:48 crc kubenswrapper[4629]: I1211 09:02:48.022935 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22d09ce1-4677-427b-8007-f8743ee036dc-catalog-content\") pod \"certified-operators-kf76h\" (UID: \"22d09ce1-4677-427b-8007-f8743ee036dc\") " pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:02:48 crc kubenswrapper[4629]: I1211 09:02:48.055181 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm5cc\" (UniqueName: \"kubernetes.io/projected/22d09ce1-4677-427b-8007-f8743ee036dc-kube-api-access-nm5cc\") pod \"certified-operators-kf76h\" (UID: \"22d09ce1-4677-427b-8007-f8743ee036dc\") " pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:02:48 crc kubenswrapper[4629]: I1211 09:02:48.117693 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:02:48 crc kubenswrapper[4629]: I1211 09:02:48.669869 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kf76h"] Dec 11 09:02:48 crc kubenswrapper[4629]: I1211 09:02:48.966576 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf76h" event={"ID":"22d09ce1-4677-427b-8007-f8743ee036dc","Type":"ContainerStarted","Data":"61f25295d30b6660896b3475f471044a92342de37b55fc5b775e744e61b20c5d"} Dec 11 09:02:49 crc kubenswrapper[4629]: I1211 09:02:49.866129 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 11 09:02:49 crc kubenswrapper[4629]: I1211 09:02:49.980203 4629 generic.go:334] "Generic (PLEG): container finished" podID="22d09ce1-4677-427b-8007-f8743ee036dc" containerID="6b00e12ea1efc899a14a19a1e9401ddcb707b0ad8f03839abe5d8f335b78e9d5" exitCode=0 Dec 11 09:02:49 crc kubenswrapper[4629]: I1211 09:02:49.980936 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf76h" event={"ID":"22d09ce1-4677-427b-8007-f8743ee036dc","Type":"ContainerDied","Data":"6b00e12ea1efc899a14a19a1e9401ddcb707b0ad8f03839abe5d8f335b78e9d5"} Dec 11 09:02:50 crc kubenswrapper[4629]: I1211 09:02:50.486040 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:02:50 crc kubenswrapper[4629]: I1211 09:02:50.486109 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:02:51 crc kubenswrapper[4629]: I1211 09:02:51.922835 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.183988 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wbqb7"] Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.189192 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.235355 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wbqb7"] Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.311001 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv5dk\" (UniqueName: \"kubernetes.io/projected/e6cae899-1710-4472-b98c-8bbbad4d3e2e-kube-api-access-qv5dk\") pod \"community-operators-wbqb7\" (UID: \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\") " pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.311158 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6cae899-1710-4472-b98c-8bbbad4d3e2e-catalog-content\") pod \"community-operators-wbqb7\" (UID: \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\") " pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.311259 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6cae899-1710-4472-b98c-8bbbad4d3e2e-utilities\") pod \"community-operators-wbqb7\" (UID: \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\") " pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.413024 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv5dk\" (UniqueName: \"kubernetes.io/projected/e6cae899-1710-4472-b98c-8bbbad4d3e2e-kube-api-access-qv5dk\") pod \"community-operators-wbqb7\" (UID: \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\") " pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.413415 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6cae899-1710-4472-b98c-8bbbad4d3e2e-catalog-content\") pod \"community-operators-wbqb7\" (UID: \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\") " pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.413465 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6cae899-1710-4472-b98c-8bbbad4d3e2e-utilities\") pod \"community-operators-wbqb7\" (UID: \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\") " pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.413944 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6cae899-1710-4472-b98c-8bbbad4d3e2e-utilities\") pod \"community-operators-wbqb7\" (UID: \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\") " pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.414195 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6cae899-1710-4472-b98c-8bbbad4d3e2e-catalog-content\") pod \"community-operators-wbqb7\" (UID: \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\") " pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.439594 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv5dk\" (UniqueName: \"kubernetes.io/projected/e6cae899-1710-4472-b98c-8bbbad4d3e2e-kube-api-access-qv5dk\") pod \"community-operators-wbqb7\" (UID: \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\") " pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:02:52 crc kubenswrapper[4629]: I1211 09:02:52.513076 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:02:53 crc kubenswrapper[4629]: I1211 09:02:53.137649 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wbqb7"] Dec 11 09:02:54 crc kubenswrapper[4629]: I1211 09:02:54.025818 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf76h" event={"ID":"22d09ce1-4677-427b-8007-f8743ee036dc","Type":"ContainerStarted","Data":"95214635a5bfdd69fee9d464ed4704bd3795741b031b77cec463e549e92ae736"} Dec 11 09:02:54 crc kubenswrapper[4629]: I1211 09:02:54.035061 4629 generic.go:334] "Generic (PLEG): container finished" podID="e6cae899-1710-4472-b98c-8bbbad4d3e2e" containerID="7922a10930f04c05077ed31db686e72349c237483cbd8679b00a579b68b02711" exitCode=0 Dec 11 09:02:54 crc kubenswrapper[4629]: I1211 09:02:54.035134 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqb7" event={"ID":"e6cae899-1710-4472-b98c-8bbbad4d3e2e","Type":"ContainerDied","Data":"7922a10930f04c05077ed31db686e72349c237483cbd8679b00a579b68b02711"} Dec 11 09:02:54 crc kubenswrapper[4629]: I1211 09:02:54.035172 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqb7" event={"ID":"e6cae899-1710-4472-b98c-8bbbad4d3e2e","Type":"ContainerStarted","Data":"40cd1c0b316a96377f8a9276393caaa25d67db48df9c99537c42b608581a66a3"} Dec 11 09:02:56 crc kubenswrapper[4629]: I1211 09:02:56.065982 4629 generic.go:334] "Generic (PLEG): container finished" podID="22d09ce1-4677-427b-8007-f8743ee036dc" containerID="95214635a5bfdd69fee9d464ed4704bd3795741b031b77cec463e549e92ae736" exitCode=0 Dec 11 09:02:56 crc kubenswrapper[4629]: I1211 09:02:56.066539 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf76h" event={"ID":"22d09ce1-4677-427b-8007-f8743ee036dc","Type":"ContainerDied","Data":"95214635a5bfdd69fee9d464ed4704bd3795741b031b77cec463e549e92ae736"} Dec 11 09:02:58 crc kubenswrapper[4629]: I1211 09:02:58.086017 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqb7" event={"ID":"e6cae899-1710-4472-b98c-8bbbad4d3e2e","Type":"ContainerStarted","Data":"ad1c3c3c6381b9c31e7e33dd0b2fc76c425743692ff56017681087e572bfcbee"} Dec 11 09:02:58 crc kubenswrapper[4629]: I1211 09:02:58.251427 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 09:02:59 crc kubenswrapper[4629]: I1211 09:02:59.096188 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" event={"ID":"0206ce20-df99-494c-96f0-f6201de98376","Type":"ContainerStarted","Data":"c23832bc6e3dcf2d0149fd953a88f7efe4e2282a748ddd0b56b3692ac3343b18"} Dec 11 09:02:59 crc kubenswrapper[4629]: I1211 09:02:59.098815 4629 generic.go:334] "Generic (PLEG): container finished" podID="e6cae899-1710-4472-b98c-8bbbad4d3e2e" containerID="ad1c3c3c6381b9c31e7e33dd0b2fc76c425743692ff56017681087e572bfcbee" exitCode=0 Dec 11 09:02:59 crc kubenswrapper[4629]: I1211 09:02:59.098868 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqb7" event={"ID":"e6cae899-1710-4472-b98c-8bbbad4d3e2e","Type":"ContainerDied","Data":"ad1c3c3c6381b9c31e7e33dd0b2fc76c425743692ff56017681087e572bfcbee"} Dec 11 09:02:59 crc kubenswrapper[4629]: I1211 09:02:59.140490 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" podStartSLOduration=2.807205707 podStartE2EDuration="37.140469517s" podCreationTimestamp="2025-12-11 09:02:22 +0000 UTC" firstStartedPulling="2025-12-11 09:02:23.915329131 +0000 UTC m=+1531.980744739" lastFinishedPulling="2025-12-11 09:02:58.248592941 +0000 UTC m=+1566.314008549" observedRunningTime="2025-12-11 09:02:59.137467583 +0000 UTC m=+1567.202883191" watchObservedRunningTime="2025-12-11 09:02:59.140469517 +0000 UTC m=+1567.205885125" Dec 11 09:03:00 crc kubenswrapper[4629]: I1211 09:03:00.114106 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqb7" event={"ID":"e6cae899-1710-4472-b98c-8bbbad4d3e2e","Type":"ContainerStarted","Data":"a4071f778afdc5a4701bc2dde7495a9a86f86574a817bca36c382710315f93f8"} Dec 11 09:03:00 crc kubenswrapper[4629]: I1211 09:03:00.138077 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf76h" event={"ID":"22d09ce1-4677-427b-8007-f8743ee036dc","Type":"ContainerStarted","Data":"ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61"} Dec 11 09:03:00 crc kubenswrapper[4629]: I1211 09:03:00.139026 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wbqb7" podStartSLOduration=2.588050581 podStartE2EDuration="8.139014285s" podCreationTimestamp="2025-12-11 09:02:52 +0000 UTC" firstStartedPulling="2025-12-11 09:02:54.03808801 +0000 UTC m=+1562.103503618" lastFinishedPulling="2025-12-11 09:02:59.589051714 +0000 UTC m=+1567.654467322" observedRunningTime="2025-12-11 09:03:00.137698453 +0000 UTC m=+1568.203114061" watchObservedRunningTime="2025-12-11 09:03:00.139014285 +0000 UTC m=+1568.204429893" Dec 11 09:03:00 crc kubenswrapper[4629]: I1211 09:03:00.172929 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kf76h" podStartSLOduration=3.740393403 podStartE2EDuration="13.172904649s" podCreationTimestamp="2025-12-11 09:02:47 +0000 UTC" firstStartedPulling="2025-12-11 09:02:49.98272617 +0000 UTC m=+1558.048141788" lastFinishedPulling="2025-12-11 09:02:59.415237426 +0000 UTC m=+1567.480653034" observedRunningTime="2025-12-11 09:03:00.1672127 +0000 UTC m=+1568.232628308" watchObservedRunningTime="2025-12-11 09:03:00.172904649 +0000 UTC m=+1568.238320257" Dec 11 09:03:02 crc kubenswrapper[4629]: I1211 09:03:02.513634 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:03:02 crc kubenswrapper[4629]: I1211 09:03:02.514014 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:03:02 crc kubenswrapper[4629]: I1211 09:03:02.560929 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:03:03 crc kubenswrapper[4629]: I1211 09:03:03.559354 4629 scope.go:117] "RemoveContainer" containerID="ddea03484aeae20a8f71069e2fc25103d6b621f967af7e32be81c909767b96a5" Dec 11 09:03:08 crc kubenswrapper[4629]: I1211 09:03:08.118151 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:03:08 crc kubenswrapper[4629]: I1211 09:03:08.118973 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:03:08 crc kubenswrapper[4629]: I1211 09:03:08.159823 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:03:08 crc kubenswrapper[4629]: I1211 09:03:08.259796 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:03:08 crc kubenswrapper[4629]: I1211 09:03:08.397063 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kf76h"] Dec 11 09:03:10 crc kubenswrapper[4629]: I1211 09:03:10.234513 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kf76h" podUID="22d09ce1-4677-427b-8007-f8743ee036dc" containerName="registry-server" containerID="cri-o://ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61" gracePeriod=2 Dec 11 09:03:10 crc kubenswrapper[4629]: I1211 09:03:10.705097 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:03:10 crc kubenswrapper[4629]: I1211 09:03:10.796567 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22d09ce1-4677-427b-8007-f8743ee036dc-catalog-content\") pod \"22d09ce1-4677-427b-8007-f8743ee036dc\" (UID: \"22d09ce1-4677-427b-8007-f8743ee036dc\") " Dec 11 09:03:10 crc kubenswrapper[4629]: I1211 09:03:10.796951 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22d09ce1-4677-427b-8007-f8743ee036dc-utilities\") pod \"22d09ce1-4677-427b-8007-f8743ee036dc\" (UID: \"22d09ce1-4677-427b-8007-f8743ee036dc\") " Dec 11 09:03:10 crc kubenswrapper[4629]: I1211 09:03:10.797135 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm5cc\" (UniqueName: \"kubernetes.io/projected/22d09ce1-4677-427b-8007-f8743ee036dc-kube-api-access-nm5cc\") pod \"22d09ce1-4677-427b-8007-f8743ee036dc\" (UID: \"22d09ce1-4677-427b-8007-f8743ee036dc\") " Dec 11 09:03:10 crc kubenswrapper[4629]: I1211 09:03:10.798697 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22d09ce1-4677-427b-8007-f8743ee036dc-utilities" (OuterVolumeSpecName: "utilities") pod "22d09ce1-4677-427b-8007-f8743ee036dc" (UID: "22d09ce1-4677-427b-8007-f8743ee036dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:03:10 crc kubenswrapper[4629]: I1211 09:03:10.803138 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22d09ce1-4677-427b-8007-f8743ee036dc-kube-api-access-nm5cc" (OuterVolumeSpecName: "kube-api-access-nm5cc") pod "22d09ce1-4677-427b-8007-f8743ee036dc" (UID: "22d09ce1-4677-427b-8007-f8743ee036dc"). InnerVolumeSpecName "kube-api-access-nm5cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:03:10 crc kubenswrapper[4629]: I1211 09:03:10.865578 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22d09ce1-4677-427b-8007-f8743ee036dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22d09ce1-4677-427b-8007-f8743ee036dc" (UID: "22d09ce1-4677-427b-8007-f8743ee036dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:03:10 crc kubenswrapper[4629]: I1211 09:03:10.898703 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22d09ce1-4677-427b-8007-f8743ee036dc-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:03:10 crc kubenswrapper[4629]: I1211 09:03:10.898747 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm5cc\" (UniqueName: \"kubernetes.io/projected/22d09ce1-4677-427b-8007-f8743ee036dc-kube-api-access-nm5cc\") on node \"crc\" DevicePath \"\"" Dec 11 09:03:10 crc kubenswrapper[4629]: I1211 09:03:10.898758 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22d09ce1-4677-427b-8007-f8743ee036dc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.246801 4629 generic.go:334] "Generic (PLEG): container finished" podID="22d09ce1-4677-427b-8007-f8743ee036dc" containerID="ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61" exitCode=0 Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.246878 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf76h" event={"ID":"22d09ce1-4677-427b-8007-f8743ee036dc","Type":"ContainerDied","Data":"ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61"} Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.246913 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf76h" event={"ID":"22d09ce1-4677-427b-8007-f8743ee036dc","Type":"ContainerDied","Data":"61f25295d30b6660896b3475f471044a92342de37b55fc5b775e744e61b20c5d"} Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.246930 4629 scope.go:117] "RemoveContainer" containerID="ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61" Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.247069 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kf76h" Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.271077 4629 scope.go:117] "RemoveContainer" containerID="95214635a5bfdd69fee9d464ed4704bd3795741b031b77cec463e549e92ae736" Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.294434 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kf76h"] Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.315140 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kf76h"] Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.324539 4629 scope.go:117] "RemoveContainer" containerID="6b00e12ea1efc899a14a19a1e9401ddcb707b0ad8f03839abe5d8f335b78e9d5" Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.346253 4629 scope.go:117] "RemoveContainer" containerID="ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61" Dec 11 09:03:11 crc kubenswrapper[4629]: E1211 09:03:11.346658 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61\": container with ID starting with ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61 not found: ID does not exist" containerID="ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61" Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.346702 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61"} err="failed to get container status \"ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61\": rpc error: code = NotFound desc = could not find container \"ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61\": container with ID starting with ad5df69d67f0444636dadf17121a9c93707117a7dd76cf5556563ae597082c61 not found: ID does not exist" Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.346731 4629 scope.go:117] "RemoveContainer" containerID="95214635a5bfdd69fee9d464ed4704bd3795741b031b77cec463e549e92ae736" Dec 11 09:03:11 crc kubenswrapper[4629]: E1211 09:03:11.347151 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95214635a5bfdd69fee9d464ed4704bd3795741b031b77cec463e549e92ae736\": container with ID starting with 95214635a5bfdd69fee9d464ed4704bd3795741b031b77cec463e549e92ae736 not found: ID does not exist" containerID="95214635a5bfdd69fee9d464ed4704bd3795741b031b77cec463e549e92ae736" Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.347177 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95214635a5bfdd69fee9d464ed4704bd3795741b031b77cec463e549e92ae736"} err="failed to get container status \"95214635a5bfdd69fee9d464ed4704bd3795741b031b77cec463e549e92ae736\": rpc error: code = NotFound desc = could not find container \"95214635a5bfdd69fee9d464ed4704bd3795741b031b77cec463e549e92ae736\": container with ID starting with 95214635a5bfdd69fee9d464ed4704bd3795741b031b77cec463e549e92ae736 not found: ID does not exist" Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.347193 4629 scope.go:117] "RemoveContainer" containerID="6b00e12ea1efc899a14a19a1e9401ddcb707b0ad8f03839abe5d8f335b78e9d5" Dec 11 09:03:11 crc kubenswrapper[4629]: E1211 09:03:11.347431 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b00e12ea1efc899a14a19a1e9401ddcb707b0ad8f03839abe5d8f335b78e9d5\": container with ID starting with 6b00e12ea1efc899a14a19a1e9401ddcb707b0ad8f03839abe5d8f335b78e9d5 not found: ID does not exist" containerID="6b00e12ea1efc899a14a19a1e9401ddcb707b0ad8f03839abe5d8f335b78e9d5" Dec 11 09:03:11 crc kubenswrapper[4629]: I1211 09:03:11.347450 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b00e12ea1efc899a14a19a1e9401ddcb707b0ad8f03839abe5d8f335b78e9d5"} err="failed to get container status \"6b00e12ea1efc899a14a19a1e9401ddcb707b0ad8f03839abe5d8f335b78e9d5\": rpc error: code = NotFound desc = could not find container \"6b00e12ea1efc899a14a19a1e9401ddcb707b0ad8f03839abe5d8f335b78e9d5\": container with ID starting with 6b00e12ea1efc899a14a19a1e9401ddcb707b0ad8f03839abe5d8f335b78e9d5 not found: ID does not exist" Dec 11 09:03:12 crc kubenswrapper[4629]: I1211 09:03:12.209813 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22d09ce1-4677-427b-8007-f8743ee036dc" path="/var/lib/kubelet/pods/22d09ce1-4677-427b-8007-f8743ee036dc/volumes" Dec 11 09:03:12 crc kubenswrapper[4629]: I1211 09:03:12.563416 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:03:13 crc kubenswrapper[4629]: I1211 09:03:13.795126 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wbqb7"] Dec 11 09:03:13 crc kubenswrapper[4629]: I1211 09:03:13.796532 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wbqb7" podUID="e6cae899-1710-4472-b98c-8bbbad4d3e2e" containerName="registry-server" containerID="cri-o://a4071f778afdc5a4701bc2dde7495a9a86f86574a817bca36c382710315f93f8" gracePeriod=2 Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.276376 4629 generic.go:334] "Generic (PLEG): container finished" podID="e6cae899-1710-4472-b98c-8bbbad4d3e2e" containerID="a4071f778afdc5a4701bc2dde7495a9a86f86574a817bca36c382710315f93f8" exitCode=0 Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.276447 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqb7" event={"ID":"e6cae899-1710-4472-b98c-8bbbad4d3e2e","Type":"ContainerDied","Data":"a4071f778afdc5a4701bc2dde7495a9a86f86574a817bca36c382710315f93f8"} Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.276739 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbqb7" event={"ID":"e6cae899-1710-4472-b98c-8bbbad4d3e2e","Type":"ContainerDied","Data":"40cd1c0b316a96377f8a9276393caaa25d67db48df9c99537c42b608581a66a3"} Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.276754 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40cd1c0b316a96377f8a9276393caaa25d67db48df9c99537c42b608581a66a3" Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.278264 4629 generic.go:334] "Generic (PLEG): container finished" podID="0206ce20-df99-494c-96f0-f6201de98376" containerID="c23832bc6e3dcf2d0149fd953a88f7efe4e2282a748ddd0b56b3692ac3343b18" exitCode=0 Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.278295 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" event={"ID":"0206ce20-df99-494c-96f0-f6201de98376","Type":"ContainerDied","Data":"c23832bc6e3dcf2d0149fd953a88f7efe4e2282a748ddd0b56b3692ac3343b18"} Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.294330 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.360538 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6cae899-1710-4472-b98c-8bbbad4d3e2e-catalog-content\") pod \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\" (UID: \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\") " Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.360611 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qv5dk\" (UniqueName: \"kubernetes.io/projected/e6cae899-1710-4472-b98c-8bbbad4d3e2e-kube-api-access-qv5dk\") pod \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\" (UID: \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\") " Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.360712 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6cae899-1710-4472-b98c-8bbbad4d3e2e-utilities\") pod \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\" (UID: \"e6cae899-1710-4472-b98c-8bbbad4d3e2e\") " Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.362414 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6cae899-1710-4472-b98c-8bbbad4d3e2e-utilities" (OuterVolumeSpecName: "utilities") pod "e6cae899-1710-4472-b98c-8bbbad4d3e2e" (UID: "e6cae899-1710-4472-b98c-8bbbad4d3e2e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.369995 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6cae899-1710-4472-b98c-8bbbad4d3e2e-kube-api-access-qv5dk" (OuterVolumeSpecName: "kube-api-access-qv5dk") pod "e6cae899-1710-4472-b98c-8bbbad4d3e2e" (UID: "e6cae899-1710-4472-b98c-8bbbad4d3e2e"). InnerVolumeSpecName "kube-api-access-qv5dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.429246 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6cae899-1710-4472-b98c-8bbbad4d3e2e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6cae899-1710-4472-b98c-8bbbad4d3e2e" (UID: "e6cae899-1710-4472-b98c-8bbbad4d3e2e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.463310 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6cae899-1710-4472-b98c-8bbbad4d3e2e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.463361 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qv5dk\" (UniqueName: \"kubernetes.io/projected/e6cae899-1710-4472-b98c-8bbbad4d3e2e-kube-api-access-qv5dk\") on node \"crc\" DevicePath \"\"" Dec 11 09:03:14 crc kubenswrapper[4629]: I1211 09:03:14.463375 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6cae899-1710-4472-b98c-8bbbad4d3e2e-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.286251 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbqb7" Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.327122 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wbqb7"] Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.337512 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wbqb7"] Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.713251 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.790694 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-ssh-key\") pod \"0206ce20-df99-494c-96f0-f6201de98376\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.791571 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-inventory\") pod \"0206ce20-df99-494c-96f0-f6201de98376\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.791658 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-repo-setup-combined-ca-bundle\") pod \"0206ce20-df99-494c-96f0-f6201de98376\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.791801 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgsgh\" (UniqueName: \"kubernetes.io/projected/0206ce20-df99-494c-96f0-f6201de98376-kube-api-access-sgsgh\") pod \"0206ce20-df99-494c-96f0-f6201de98376\" (UID: \"0206ce20-df99-494c-96f0-f6201de98376\") " Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.795878 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "0206ce20-df99-494c-96f0-f6201de98376" (UID: "0206ce20-df99-494c-96f0-f6201de98376"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.797891 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0206ce20-df99-494c-96f0-f6201de98376-kube-api-access-sgsgh" (OuterVolumeSpecName: "kube-api-access-sgsgh") pod "0206ce20-df99-494c-96f0-f6201de98376" (UID: "0206ce20-df99-494c-96f0-f6201de98376"). InnerVolumeSpecName "kube-api-access-sgsgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.820038 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-inventory" (OuterVolumeSpecName: "inventory") pod "0206ce20-df99-494c-96f0-f6201de98376" (UID: "0206ce20-df99-494c-96f0-f6201de98376"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.820378 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0206ce20-df99-494c-96f0-f6201de98376" (UID: "0206ce20-df99-494c-96f0-f6201de98376"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.893863 4629 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.893919 4629 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.893931 4629 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0206ce20-df99-494c-96f0-f6201de98376-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:03:15 crc kubenswrapper[4629]: I1211 09:03:15.893943 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgsgh\" (UniqueName: \"kubernetes.io/projected/0206ce20-df99-494c-96f0-f6201de98376-kube-api-access-sgsgh\") on node \"crc\" DevicePath \"\"" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.209770 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6cae899-1710-4472-b98c-8bbbad4d3e2e" path="/var/lib/kubelet/pods/e6cae899-1710-4472-b98c-8bbbad4d3e2e/volumes" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.295593 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" event={"ID":"0206ce20-df99-494c-96f0-f6201de98376","Type":"ContainerDied","Data":"85278d128cb805fbd31666e91c64e269fe8dd6413ed5bb8e6b8a6e11a7a4744b"} Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.295637 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85278d128cb805fbd31666e91c64e269fe8dd6413ed5bb8e6b8a6e11a7a4744b" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.295719 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.385226 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w"] Dec 11 09:03:16 crc kubenswrapper[4629]: E1211 09:03:16.386014 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6cae899-1710-4472-b98c-8bbbad4d3e2e" containerName="registry-server" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.386032 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6cae899-1710-4472-b98c-8bbbad4d3e2e" containerName="registry-server" Dec 11 09:03:16 crc kubenswrapper[4629]: E1211 09:03:16.386047 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6cae899-1710-4472-b98c-8bbbad4d3e2e" containerName="extract-content" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.386053 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6cae899-1710-4472-b98c-8bbbad4d3e2e" containerName="extract-content" Dec 11 09:03:16 crc kubenswrapper[4629]: E1211 09:03:16.386067 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22d09ce1-4677-427b-8007-f8743ee036dc" containerName="extract-utilities" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.386075 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="22d09ce1-4677-427b-8007-f8743ee036dc" containerName="extract-utilities" Dec 11 09:03:16 crc kubenswrapper[4629]: E1211 09:03:16.386086 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0206ce20-df99-494c-96f0-f6201de98376" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.386093 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="0206ce20-df99-494c-96f0-f6201de98376" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 11 09:03:16 crc kubenswrapper[4629]: E1211 09:03:16.386112 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22d09ce1-4677-427b-8007-f8743ee036dc" containerName="extract-content" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.386118 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="22d09ce1-4677-427b-8007-f8743ee036dc" containerName="extract-content" Dec 11 09:03:16 crc kubenswrapper[4629]: E1211 09:03:16.386129 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22d09ce1-4677-427b-8007-f8743ee036dc" containerName="registry-server" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.386143 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="22d09ce1-4677-427b-8007-f8743ee036dc" containerName="registry-server" Dec 11 09:03:16 crc kubenswrapper[4629]: E1211 09:03:16.386152 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6cae899-1710-4472-b98c-8bbbad4d3e2e" containerName="extract-utilities" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.386158 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6cae899-1710-4472-b98c-8bbbad4d3e2e" containerName="extract-utilities" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.386329 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6cae899-1710-4472-b98c-8bbbad4d3e2e" containerName="registry-server" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.386351 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="0206ce20-df99-494c-96f0-f6201de98376" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.386370 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="22d09ce1-4677-427b-8007-f8743ee036dc" containerName="registry-server" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.387029 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.390175 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qjfc4" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.390371 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.403899 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.404181 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.406399 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w"] Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.514954 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.515070 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.515100 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.515164 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7fln\" (UniqueName: \"kubernetes.io/projected/d898802e-842a-468e-a468-8aa61c2c9360-kube-api-access-f7fln\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.616705 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.616832 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.616919 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.617036 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7fln\" (UniqueName: \"kubernetes.io/projected/d898802e-842a-468e-a468-8aa61c2c9360-kube-api-access-f7fln\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.621516 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.631390 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.633270 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.637265 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7fln\" (UniqueName: \"kubernetes.io/projected/d898802e-842a-468e-a468-8aa61c2c9360-kube-api-access-f7fln\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:16 crc kubenswrapper[4629]: I1211 09:03:16.704224 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:03:17 crc kubenswrapper[4629]: I1211 09:03:17.275481 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w"] Dec 11 09:03:17 crc kubenswrapper[4629]: I1211 09:03:17.307223 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" event={"ID":"d898802e-842a-468e-a468-8aa61c2c9360","Type":"ContainerStarted","Data":"a4e49f2cfbb0374939db720593705cddbb754580728149dad1c85b3d443c1837"} Dec 11 09:03:18 crc kubenswrapper[4629]: I1211 09:03:18.316137 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" event={"ID":"d898802e-842a-468e-a468-8aa61c2c9360","Type":"ContainerStarted","Data":"b2493398de24985ea66df136bd4d1d25b6c58aae12b1a0d7f359d9d4db225ed4"} Dec 11 09:03:18 crc kubenswrapper[4629]: I1211 09:03:18.338981 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" podStartSLOduration=1.802018446 podStartE2EDuration="2.338758511s" podCreationTimestamp="2025-12-11 09:03:16 +0000 UTC" firstStartedPulling="2025-12-11 09:03:17.284810014 +0000 UTC m=+1585.350225622" lastFinishedPulling="2025-12-11 09:03:17.821550079 +0000 UTC m=+1585.886965687" observedRunningTime="2025-12-11 09:03:18.332470973 +0000 UTC m=+1586.397886581" watchObservedRunningTime="2025-12-11 09:03:18.338758511 +0000 UTC m=+1586.404174119" Dec 11 09:03:20 crc kubenswrapper[4629]: I1211 09:03:20.485922 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:03:20 crc kubenswrapper[4629]: I1211 09:03:20.487190 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:03:50 crc kubenswrapper[4629]: I1211 09:03:50.487334 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:03:50 crc kubenswrapper[4629]: I1211 09:03:50.488322 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:03:50 crc kubenswrapper[4629]: I1211 09:03:50.488397 4629 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 09:03:50 crc kubenswrapper[4629]: I1211 09:03:50.489463 4629 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f"} pod="openshift-machine-config-operator/machine-config-daemon-wbld9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 09:03:50 crc kubenswrapper[4629]: I1211 09:03:50.489535 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" containerID="cri-o://099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" gracePeriod=600 Dec 11 09:03:51 crc kubenswrapper[4629]: E1211 09:03:51.122706 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:03:51 crc kubenswrapper[4629]: I1211 09:03:51.616546 4629 generic.go:334] "Generic (PLEG): container finished" podID="847b50b9-0093-404f-b0dd-63c814d8986b" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" exitCode=0 Dec 11 09:03:51 crc kubenswrapper[4629]: I1211 09:03:51.616917 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerDied","Data":"099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f"} Dec 11 09:03:51 crc kubenswrapper[4629]: I1211 09:03:51.616950 4629 scope.go:117] "RemoveContainer" containerID="f34698209a804ac13c8c04a3199ac3381f69055cef9be6fa18f868fa21af6591" Dec 11 09:03:51 crc kubenswrapper[4629]: I1211 09:03:51.617550 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:03:51 crc kubenswrapper[4629]: E1211 09:03:51.617799 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:04:03 crc kubenswrapper[4629]: I1211 09:04:03.692428 4629 scope.go:117] "RemoveContainer" containerID="44d6f0deeece43bb1b79e87e7a318cc1e99e9f992cb9d9bf865246d992cc0b60" Dec 11 09:04:03 crc kubenswrapper[4629]: I1211 09:04:03.719508 4629 scope.go:117] "RemoveContainer" containerID="a03460dd9d0276ab2a93e7fc6d9666994f9e74e8276b474ef8cdc3c6026bfb01" Dec 11 09:04:03 crc kubenswrapper[4629]: I1211 09:04:03.740667 4629 scope.go:117] "RemoveContainer" containerID="fabd10859759368fe09db4183b04a74db915ffa14133f4e49319ebc339cc439c" Dec 11 09:04:04 crc kubenswrapper[4629]: I1211 09:04:04.199432 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:04:04 crc kubenswrapper[4629]: E1211 09:04:04.200063 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:04:15 crc kubenswrapper[4629]: I1211 09:04:15.199705 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:04:15 crc kubenswrapper[4629]: E1211 09:04:15.201292 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:04:28 crc kubenswrapper[4629]: I1211 09:04:28.199055 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:04:28 crc kubenswrapper[4629]: E1211 09:04:28.199969 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:04:39 crc kubenswrapper[4629]: I1211 09:04:39.199500 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:04:39 crc kubenswrapper[4629]: E1211 09:04:39.200373 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:04:54 crc kubenswrapper[4629]: I1211 09:04:54.199057 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:04:54 crc kubenswrapper[4629]: E1211 09:04:54.199901 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:05:03 crc kubenswrapper[4629]: I1211 09:05:03.833237 4629 scope.go:117] "RemoveContainer" containerID="d9806635827801334b3adc8ba82a4e5d638d7180d14e191e460b9a56999e7df5" Dec 11 09:05:03 crc kubenswrapper[4629]: I1211 09:05:03.876828 4629 scope.go:117] "RemoveContainer" containerID="185de34a4d8206be0d10ab57a08a2789aa2e36b035d5919e8b467721eb186166" Dec 11 09:05:03 crc kubenswrapper[4629]: I1211 09:05:03.904268 4629 scope.go:117] "RemoveContainer" containerID="c5d02d1f287d826d0f0b38903c19ddc5a50e9b6ed886049f20918751c41c5e1e" Dec 11 09:05:08 crc kubenswrapper[4629]: I1211 09:05:08.199322 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:05:08 crc kubenswrapper[4629]: E1211 09:05:08.200633 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:05:21 crc kubenswrapper[4629]: I1211 09:05:21.199753 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:05:21 crc kubenswrapper[4629]: E1211 09:05:21.200867 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:05:36 crc kubenswrapper[4629]: I1211 09:05:36.199439 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:05:36 crc kubenswrapper[4629]: E1211 09:05:36.200290 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:05:47 crc kubenswrapper[4629]: I1211 09:05:47.199415 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:05:47 crc kubenswrapper[4629]: E1211 09:05:47.200248 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:06:01 crc kubenswrapper[4629]: I1211 09:06:01.199433 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:06:01 crc kubenswrapper[4629]: E1211 09:06:01.200314 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:06:14 crc kubenswrapper[4629]: I1211 09:06:14.199632 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:06:14 crc kubenswrapper[4629]: E1211 09:06:14.200670 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:06:29 crc kubenswrapper[4629]: I1211 09:06:29.199828 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:06:29 crc kubenswrapper[4629]: E1211 09:06:29.200590 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:06:34 crc kubenswrapper[4629]: I1211 09:06:34.064955 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-tzqff"] Dec 11 09:06:34 crc kubenswrapper[4629]: I1211 09:06:34.073947 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-005f-account-create-update-d7fhm"] Dec 11 09:06:34 crc kubenswrapper[4629]: I1211 09:06:34.082269 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-005f-account-create-update-d7fhm"] Dec 11 09:06:34 crc kubenswrapper[4629]: I1211 09:06:34.090278 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-tzqff"] Dec 11 09:06:34 crc kubenswrapper[4629]: I1211 09:06:34.208236 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48b7b556-b31c-4a94-aa25-3a26ba8d07dd" path="/var/lib/kubelet/pods/48b7b556-b31c-4a94-aa25-3a26ba8d07dd/volumes" Dec 11 09:06:34 crc kubenswrapper[4629]: I1211 09:06:34.209175 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54cce582-17c9-40d7-92f1-8a7c05a8ca57" path="/var/lib/kubelet/pods/54cce582-17c9-40d7-92f1-8a7c05a8ca57/volumes" Dec 11 09:06:36 crc kubenswrapper[4629]: I1211 09:06:36.032508 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-2sknt"] Dec 11 09:06:36 crc kubenswrapper[4629]: I1211 09:06:36.039382 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a1e4-account-create-update-rlcdj"] Dec 11 09:06:36 crc kubenswrapper[4629]: I1211 09:06:36.050531 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-grhlk"] Dec 11 09:06:36 crc kubenswrapper[4629]: I1211 09:06:36.080034 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-8495-account-create-update-5rsvg"] Dec 11 09:06:36 crc kubenswrapper[4629]: I1211 09:06:36.091510 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a1e4-account-create-update-rlcdj"] Dec 11 09:06:36 crc kubenswrapper[4629]: I1211 09:06:36.099558 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-8495-account-create-update-5rsvg"] Dec 11 09:06:36 crc kubenswrapper[4629]: I1211 09:06:36.109736 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-2sknt"] Dec 11 09:06:36 crc kubenswrapper[4629]: I1211 09:06:36.121435 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-grhlk"] Dec 11 09:06:36 crc kubenswrapper[4629]: I1211 09:06:36.210691 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e81ddaf-9300-4a5b-b3cf-5f90928921a2" path="/var/lib/kubelet/pods/0e81ddaf-9300-4a5b-b3cf-5f90928921a2/volumes" Dec 11 09:06:36 crc kubenswrapper[4629]: I1211 09:06:36.211762 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3da40552-275b-4c71-a2d5-2257bdad664d" path="/var/lib/kubelet/pods/3da40552-275b-4c71-a2d5-2257bdad664d/volumes" Dec 11 09:06:36 crc kubenswrapper[4629]: I1211 09:06:36.212551 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="623e0d38-cbc4-4261-ad37-319eb1704ce2" path="/var/lib/kubelet/pods/623e0d38-cbc4-4261-ad37-319eb1704ce2/volumes" Dec 11 09:06:36 crc kubenswrapper[4629]: I1211 09:06:36.213307 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7f3b384-3bf7-46e1-9316-9c83bac9a8ac" path="/var/lib/kubelet/pods/b7f3b384-3bf7-46e1-9316-9c83bac9a8ac/volumes" Dec 11 09:06:40 crc kubenswrapper[4629]: I1211 09:06:40.200329 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:06:40 crc kubenswrapper[4629]: E1211 09:06:40.200837 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:06:53 crc kubenswrapper[4629]: I1211 09:06:53.199039 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:06:53 crc kubenswrapper[4629]: E1211 09:06:53.201416 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:06:59 crc kubenswrapper[4629]: I1211 09:06:59.596956 4629 generic.go:334] "Generic (PLEG): container finished" podID="d898802e-842a-468e-a468-8aa61c2c9360" containerID="b2493398de24985ea66df136bd4d1d25b6c58aae12b1a0d7f359d9d4db225ed4" exitCode=0 Dec 11 09:06:59 crc kubenswrapper[4629]: I1211 09:06:59.597134 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" event={"ID":"d898802e-842a-468e-a468-8aa61c2c9360","Type":"ContainerDied","Data":"b2493398de24985ea66df136bd4d1d25b6c58aae12b1a0d7f359d9d4db225ed4"} Dec 11 09:07:00 crc kubenswrapper[4629]: I1211 09:07:00.988749 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.067257 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-ssh-key\") pod \"d898802e-842a-468e-a468-8aa61c2c9360\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.067509 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7fln\" (UniqueName: \"kubernetes.io/projected/d898802e-842a-468e-a468-8aa61c2c9360-kube-api-access-f7fln\") pod \"d898802e-842a-468e-a468-8aa61c2c9360\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.067568 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-inventory\") pod \"d898802e-842a-468e-a468-8aa61c2c9360\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.067594 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-bootstrap-combined-ca-bundle\") pod \"d898802e-842a-468e-a468-8aa61c2c9360\" (UID: \"d898802e-842a-468e-a468-8aa61c2c9360\") " Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.093912 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d898802e-842a-468e-a468-8aa61c2c9360-kube-api-access-f7fln" (OuterVolumeSpecName: "kube-api-access-f7fln") pod "d898802e-842a-468e-a468-8aa61c2c9360" (UID: "d898802e-842a-468e-a468-8aa61c2c9360"). InnerVolumeSpecName "kube-api-access-f7fln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.094270 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d898802e-842a-468e-a468-8aa61c2c9360" (UID: "d898802e-842a-468e-a468-8aa61c2c9360"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.126882 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-inventory" (OuterVolumeSpecName: "inventory") pod "d898802e-842a-468e-a468-8aa61c2c9360" (UID: "d898802e-842a-468e-a468-8aa61c2c9360"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.162005 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d898802e-842a-468e-a468-8aa61c2c9360" (UID: "d898802e-842a-468e-a468-8aa61c2c9360"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.170134 4629 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.170173 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7fln\" (UniqueName: \"kubernetes.io/projected/d898802e-842a-468e-a468-8aa61c2c9360-kube-api-access-f7fln\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.170188 4629 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.170198 4629 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d898802e-842a-468e-a468-8aa61c2c9360-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.616371 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" event={"ID":"d898802e-842a-468e-a468-8aa61c2c9360","Type":"ContainerDied","Data":"a4e49f2cfbb0374939db720593705cddbb754580728149dad1c85b3d443c1837"} Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.616406 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.616432 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4e49f2cfbb0374939db720593705cddbb754580728149dad1c85b3d443c1837" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.703090 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq"] Dec 11 09:07:01 crc kubenswrapper[4629]: E1211 09:07:01.703781 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d898802e-842a-468e-a468-8aa61c2c9360" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.703890 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="d898802e-842a-468e-a468-8aa61c2c9360" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.704172 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="d898802e-842a-468e-a468-8aa61c2c9360" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.704944 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.709888 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.718899 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq"] Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.720451 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.720888 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.721562 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qjfc4" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.779891 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea77da00-5dbb-4f8a-883c-3c8984c55122-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq\" (UID: \"ea77da00-5dbb-4f8a-883c-3c8984c55122\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.779986 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea77da00-5dbb-4f8a-883c-3c8984c55122-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq\" (UID: \"ea77da00-5dbb-4f8a-883c-3c8984c55122\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.780015 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mbw2\" (UniqueName: \"kubernetes.io/projected/ea77da00-5dbb-4f8a-883c-3c8984c55122-kube-api-access-4mbw2\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq\" (UID: \"ea77da00-5dbb-4f8a-883c-3c8984c55122\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.882442 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea77da00-5dbb-4f8a-883c-3c8984c55122-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq\" (UID: \"ea77da00-5dbb-4f8a-883c-3c8984c55122\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.882877 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea77da00-5dbb-4f8a-883c-3c8984c55122-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq\" (UID: \"ea77da00-5dbb-4f8a-883c-3c8984c55122\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.882906 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mbw2\" (UniqueName: \"kubernetes.io/projected/ea77da00-5dbb-4f8a-883c-3c8984c55122-kube-api-access-4mbw2\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq\" (UID: \"ea77da00-5dbb-4f8a-883c-3c8984c55122\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.888456 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea77da00-5dbb-4f8a-883c-3c8984c55122-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq\" (UID: \"ea77da00-5dbb-4f8a-883c-3c8984c55122\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.889442 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea77da00-5dbb-4f8a-883c-3c8984c55122-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq\" (UID: \"ea77da00-5dbb-4f8a-883c-3c8984c55122\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:07:01 crc kubenswrapper[4629]: I1211 09:07:01.897754 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mbw2\" (UniqueName: \"kubernetes.io/projected/ea77da00-5dbb-4f8a-883c-3c8984c55122-kube-api-access-4mbw2\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq\" (UID: \"ea77da00-5dbb-4f8a-883c-3c8984c55122\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:07:02 crc kubenswrapper[4629]: I1211 09:07:02.023802 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:07:02 crc kubenswrapper[4629]: I1211 09:07:02.572354 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq"] Dec 11 09:07:02 crc kubenswrapper[4629]: W1211 09:07:02.587886 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea77da00_5dbb_4f8a_883c_3c8984c55122.slice/crio-473be56af181f2813044f22d6a8acfd77c39b590221661c1f63259b07189ec27 WatchSource:0}: Error finding container 473be56af181f2813044f22d6a8acfd77c39b590221661c1f63259b07189ec27: Status 404 returned error can't find the container with id 473be56af181f2813044f22d6a8acfd77c39b590221661c1f63259b07189ec27 Dec 11 09:07:02 crc kubenswrapper[4629]: I1211 09:07:02.590751 4629 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 09:07:02 crc kubenswrapper[4629]: I1211 09:07:02.627438 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" event={"ID":"ea77da00-5dbb-4f8a-883c-3c8984c55122","Type":"ContainerStarted","Data":"473be56af181f2813044f22d6a8acfd77c39b590221661c1f63259b07189ec27"} Dec 11 09:07:03 crc kubenswrapper[4629]: I1211 09:07:03.647516 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" event={"ID":"ea77da00-5dbb-4f8a-883c-3c8984c55122","Type":"ContainerStarted","Data":"32b7bb1ca5f4cde608b79d30bdc0a2ae2d70d59657701c85d2f59c940e8ade0b"} Dec 11 09:07:03 crc kubenswrapper[4629]: I1211 09:07:03.684803 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" podStartSLOduration=2.101425542 podStartE2EDuration="2.684743769s" podCreationTimestamp="2025-12-11 09:07:01 +0000 UTC" firstStartedPulling="2025-12-11 09:07:02.590449066 +0000 UTC m=+1810.655864674" lastFinishedPulling="2025-12-11 09:07:03.173767293 +0000 UTC m=+1811.239182901" observedRunningTime="2025-12-11 09:07:03.671810393 +0000 UTC m=+1811.737226001" watchObservedRunningTime="2025-12-11 09:07:03.684743769 +0000 UTC m=+1811.750159377" Dec 11 09:07:04 crc kubenswrapper[4629]: I1211 09:07:04.014657 4629 scope.go:117] "RemoveContainer" containerID="d3059bb274a5022af71e9262aa740e8c6cb9ad05b76b83e91a430312f6cca7e7" Dec 11 09:07:04 crc kubenswrapper[4629]: I1211 09:07:04.040967 4629 scope.go:117] "RemoveContainer" containerID="3c12541e46b5d0ab526e33aa06343d192ff389f9853811ffd416d091b4ac8a76" Dec 11 09:07:04 crc kubenswrapper[4629]: I1211 09:07:04.096396 4629 scope.go:117] "RemoveContainer" containerID="0f16e1dcaba3fa2a7186d37a228a9d8fa0e4bddc8c6c74afa977e27d872a7fe7" Dec 11 09:07:04 crc kubenswrapper[4629]: I1211 09:07:04.132481 4629 scope.go:117] "RemoveContainer" containerID="2ed85de25378bb5b1589ce34a2970a8597348dce074d430820506e21f5e1f488" Dec 11 09:07:04 crc kubenswrapper[4629]: I1211 09:07:04.154729 4629 scope.go:117] "RemoveContainer" containerID="8cdd2e499d8a8d46810c4f6a3b338bc25758cb8545afb5aca32df8a634eabb75" Dec 11 09:07:04 crc kubenswrapper[4629]: I1211 09:07:04.173703 4629 scope.go:117] "RemoveContainer" containerID="a107b196af4a4e5d373352c6c1f7fa8cc040a255a4a8eb20d4d7d4545d1e32a6" Dec 11 09:07:04 crc kubenswrapper[4629]: I1211 09:07:04.199612 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:07:04 crc kubenswrapper[4629]: E1211 09:07:04.199901 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:07:04 crc kubenswrapper[4629]: I1211 09:07:04.238479 4629 scope.go:117] "RemoveContainer" containerID="7fc7ce5deccc9a06d29ba025efbe3ef62969d3f1374f909f587815d813d69aa8" Dec 11 09:07:04 crc kubenswrapper[4629]: I1211 09:07:04.276975 4629 scope.go:117] "RemoveContainer" containerID="5f1bc94f5c7ffcf73fd858a15f2e7f06042ad0d3eceffee0499845e2b0a78215" Dec 11 09:07:05 crc kubenswrapper[4629]: I1211 09:07:05.041200 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-ww7bz"] Dec 11 09:07:05 crc kubenswrapper[4629]: I1211 09:07:05.051596 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-fxcjb"] Dec 11 09:07:05 crc kubenswrapper[4629]: I1211 09:07:05.062080 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-ww7bz"] Dec 11 09:07:05 crc kubenswrapper[4629]: I1211 09:07:05.069728 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-fxcjb"] Dec 11 09:07:06 crc kubenswrapper[4629]: I1211 09:07:06.208927 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ed3ed74-d821-4428-b47b-50fdc38d6bf2" path="/var/lib/kubelet/pods/4ed3ed74-d821-4428-b47b-50fdc38d6bf2/volumes" Dec 11 09:07:06 crc kubenswrapper[4629]: I1211 09:07:06.210376 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac490e45-b339-4803-b79e-075da951724b" path="/var/lib/kubelet/pods/ac490e45-b339-4803-b79e-075da951724b/volumes" Dec 11 09:07:07 crc kubenswrapper[4629]: I1211 09:07:07.035255 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-d2wsq"] Dec 11 09:07:07 crc kubenswrapper[4629]: I1211 09:07:07.050976 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-d2wsq"] Dec 11 09:07:08 crc kubenswrapper[4629]: I1211 09:07:08.036942 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-df7c-account-create-update-5f4c6"] Dec 11 09:07:08 crc kubenswrapper[4629]: I1211 09:07:08.052960 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-4a47-account-create-update-qbfpz"] Dec 11 09:07:08 crc kubenswrapper[4629]: I1211 09:07:08.060276 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-df7c-account-create-update-5f4c6"] Dec 11 09:07:08 crc kubenswrapper[4629]: I1211 09:07:08.101920 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-0b3c-account-create-update-7wnbl"] Dec 11 09:07:08 crc kubenswrapper[4629]: I1211 09:07:08.110182 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-4a47-account-create-update-qbfpz"] Dec 11 09:07:08 crc kubenswrapper[4629]: I1211 09:07:08.118200 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-0b3c-account-create-update-7wnbl"] Dec 11 09:07:08 crc kubenswrapper[4629]: I1211 09:07:08.214377 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="161bc45b-0f32-488e-bda2-eb150afafcda" path="/var/lib/kubelet/pods/161bc45b-0f32-488e-bda2-eb150afafcda/volumes" Dec 11 09:07:08 crc kubenswrapper[4629]: I1211 09:07:08.215001 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="221a021f-7c38-4604-b717-2a93af571c05" path="/var/lib/kubelet/pods/221a021f-7c38-4604-b717-2a93af571c05/volumes" Dec 11 09:07:08 crc kubenswrapper[4629]: I1211 09:07:08.215621 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71654a6c-0437-494a-9162-a1dcd1d3547f" path="/var/lib/kubelet/pods/71654a6c-0437-494a-9162-a1dcd1d3547f/volumes" Dec 11 09:07:08 crc kubenswrapper[4629]: I1211 09:07:08.216193 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b387bcfc-8cc5-42ac-a956-4d890b023403" path="/var/lib/kubelet/pods/b387bcfc-8cc5-42ac-a956-4d890b023403/volumes" Dec 11 09:07:15 crc kubenswrapper[4629]: I1211 09:07:15.198565 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:07:15 crc kubenswrapper[4629]: E1211 09:07:15.199377 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:07:25 crc kubenswrapper[4629]: I1211 09:07:25.065536 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-gvk6g"] Dec 11 09:07:25 crc kubenswrapper[4629]: I1211 09:07:25.074935 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-gvk6g"] Dec 11 09:07:26 crc kubenswrapper[4629]: I1211 09:07:26.201251 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:07:26 crc kubenswrapper[4629]: E1211 09:07:26.202113 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:07:26 crc kubenswrapper[4629]: I1211 09:07:26.213453 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29518fb5-bde7-4468-860f-37b425465f70" path="/var/lib/kubelet/pods/29518fb5-bde7-4468-860f-37b425465f70/volumes" Dec 11 09:07:33 crc kubenswrapper[4629]: I1211 09:07:33.027189 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-j8tjt"] Dec 11 09:07:33 crc kubenswrapper[4629]: I1211 09:07:33.036440 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-j8tjt"] Dec 11 09:07:34 crc kubenswrapper[4629]: I1211 09:07:34.210187 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c469e9bc-a78a-41cc-8034-1fbec2d2452e" path="/var/lib/kubelet/pods/c469e9bc-a78a-41cc-8034-1fbec2d2452e/volumes" Dec 11 09:07:37 crc kubenswrapper[4629]: I1211 09:07:37.198999 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:07:37 crc kubenswrapper[4629]: E1211 09:07:37.199725 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:07:52 crc kubenswrapper[4629]: I1211 09:07:52.209022 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:07:52 crc kubenswrapper[4629]: E1211 09:07:52.211244 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:08:02 crc kubenswrapper[4629]: I1211 09:08:02.104116 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-4sgng"] Dec 11 09:08:02 crc kubenswrapper[4629]: I1211 09:08:02.115867 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-4sgng"] Dec 11 09:08:02 crc kubenswrapper[4629]: I1211 09:08:02.208697 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cd464f4-641f-44e9-aebd-8e4766e6185a" path="/var/lib/kubelet/pods/1cd464f4-641f-44e9-aebd-8e4766e6185a/volumes" Dec 11 09:08:03 crc kubenswrapper[4629]: I1211 09:08:03.031137 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2thgx"] Dec 11 09:08:03 crc kubenswrapper[4629]: I1211 09:08:03.038948 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2thgx"] Dec 11 09:08:03 crc kubenswrapper[4629]: I1211 09:08:03.050491 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-fb8kh"] Dec 11 09:08:03 crc kubenswrapper[4629]: I1211 09:08:03.061974 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-fb8kh"] Dec 11 09:08:03 crc kubenswrapper[4629]: I1211 09:08:03.199022 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:08:03 crc kubenswrapper[4629]: E1211 09:08:03.199269 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.211001 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b3da647-a18c-4672-9757-cbeab4f49fe0" path="/var/lib/kubelet/pods/6b3da647-a18c-4672-9757-cbeab4f49fe0/volumes" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.229110 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23" path="/var/lib/kubelet/pods/ff5e5cde-7c6c-4e5a-bfc8-c498c76ade23/volumes" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.460428 4629 scope.go:117] "RemoveContainer" containerID="40645b8724f7042c62ffe3075af810dd6c88b2a6eb11e7dbc43ec1b2880b318f" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.482452 4629 scope.go:117] "RemoveContainer" containerID="986b110b28a51ad0ab15b39f0509c6e1fee8d6c32c251fb63fee8c4d441b78f9" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.547505 4629 scope.go:117] "RemoveContainer" containerID="6dfe4c4c6a971a1a0a70a76875dd8e775ec48a6d91dd53f9cbf82c29bc0a250e" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.591089 4629 scope.go:117] "RemoveContainer" containerID="941118659c828ad5c0454bbf7f8f880df80ff439fed7cbc1eaa72e36101fdb27" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.632321 4629 scope.go:117] "RemoveContainer" containerID="477b0a96d0245e344b8d39c2a0d564c7dc703a5c8a8fcc462dce9d2e2cc4ae49" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.685488 4629 scope.go:117] "RemoveContainer" containerID="6494e783f8ad29f5503fa3c93dd38053454239eb9cf7e0d97c1c237f3efab794" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.715555 4629 scope.go:117] "RemoveContainer" containerID="04432ad6ca030050521700b021ec0a7a68bd5c5b0ce9bc42f01293244144941a" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.734001 4629 scope.go:117] "RemoveContainer" containerID="74e330e1f6221178db4c4901a82fdcbe28429b4d43fade140f4657a86594d26f" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.774724 4629 scope.go:117] "RemoveContainer" containerID="643e2f4f71712537960ccc40d9b0caa65a1eb2cfcf109ca3fbb44aa1143b6efb" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.811348 4629 scope.go:117] "RemoveContainer" containerID="98634e21d056b976e6f7d6b62c7ffa718ce74be85a1573ba4491002e5e5a3eb7" Dec 11 09:08:04 crc kubenswrapper[4629]: I1211 09:08:04.830181 4629 scope.go:117] "RemoveContainer" containerID="851281ce78b0aadd6cf06383ef11bf1bf1d2b7d105886948b3b2bb5e499bef2a" Dec 11 09:08:16 crc kubenswrapper[4629]: I1211 09:08:16.200201 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:08:16 crc kubenswrapper[4629]: E1211 09:08:16.201108 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:08:23 crc kubenswrapper[4629]: I1211 09:08:23.453121 4629 generic.go:334] "Generic (PLEG): container finished" podID="ea77da00-5dbb-4f8a-883c-3c8984c55122" containerID="32b7bb1ca5f4cde608b79d30bdc0a2ae2d70d59657701c85d2f59c940e8ade0b" exitCode=0 Dec 11 09:08:23 crc kubenswrapper[4629]: I1211 09:08:23.453193 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" event={"ID":"ea77da00-5dbb-4f8a-883c-3c8984c55122","Type":"ContainerDied","Data":"32b7bb1ca5f4cde608b79d30bdc0a2ae2d70d59657701c85d2f59c940e8ade0b"} Dec 11 09:08:24 crc kubenswrapper[4629]: I1211 09:08:24.042979 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-pdwnh"] Dec 11 09:08:24 crc kubenswrapper[4629]: I1211 09:08:24.052467 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-pdwnh"] Dec 11 09:08:24 crc kubenswrapper[4629]: I1211 09:08:24.210124 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c198d7a3-15c2-4716-b9a2-4d008076bfdb" path="/var/lib/kubelet/pods/c198d7a3-15c2-4716-b9a2-4d008076bfdb/volumes" Dec 11 09:08:24 crc kubenswrapper[4629]: I1211 09:08:24.922872 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:08:24 crc kubenswrapper[4629]: I1211 09:08:24.994101 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea77da00-5dbb-4f8a-883c-3c8984c55122-inventory\") pod \"ea77da00-5dbb-4f8a-883c-3c8984c55122\" (UID: \"ea77da00-5dbb-4f8a-883c-3c8984c55122\") " Dec 11 09:08:24 crc kubenswrapper[4629]: I1211 09:08:24.994164 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mbw2\" (UniqueName: \"kubernetes.io/projected/ea77da00-5dbb-4f8a-883c-3c8984c55122-kube-api-access-4mbw2\") pod \"ea77da00-5dbb-4f8a-883c-3c8984c55122\" (UID: \"ea77da00-5dbb-4f8a-883c-3c8984c55122\") " Dec 11 09:08:24 crc kubenswrapper[4629]: I1211 09:08:24.994271 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea77da00-5dbb-4f8a-883c-3c8984c55122-ssh-key\") pod \"ea77da00-5dbb-4f8a-883c-3c8984c55122\" (UID: \"ea77da00-5dbb-4f8a-883c-3c8984c55122\") " Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.000322 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea77da00-5dbb-4f8a-883c-3c8984c55122-kube-api-access-4mbw2" (OuterVolumeSpecName: "kube-api-access-4mbw2") pod "ea77da00-5dbb-4f8a-883c-3c8984c55122" (UID: "ea77da00-5dbb-4f8a-883c-3c8984c55122"). InnerVolumeSpecName "kube-api-access-4mbw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.028170 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea77da00-5dbb-4f8a-883c-3c8984c55122-inventory" (OuterVolumeSpecName: "inventory") pod "ea77da00-5dbb-4f8a-883c-3c8984c55122" (UID: "ea77da00-5dbb-4f8a-883c-3c8984c55122"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.036079 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea77da00-5dbb-4f8a-883c-3c8984c55122-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ea77da00-5dbb-4f8a-883c-3c8984c55122" (UID: "ea77da00-5dbb-4f8a-883c-3c8984c55122"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.096810 4629 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea77da00-5dbb-4f8a-883c-3c8984c55122-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.097108 4629 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea77da00-5dbb-4f8a-883c-3c8984c55122-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.097172 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mbw2\" (UniqueName: \"kubernetes.io/projected/ea77da00-5dbb-4f8a-883c-3c8984c55122-kube-api-access-4mbw2\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.473869 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" event={"ID":"ea77da00-5dbb-4f8a-883c-3c8984c55122","Type":"ContainerDied","Data":"473be56af181f2813044f22d6a8acfd77c39b590221661c1f63259b07189ec27"} Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.473911 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="473be56af181f2813044f22d6a8acfd77c39b590221661c1f63259b07189ec27" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.473947 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.562808 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt"] Dec 11 09:08:25 crc kubenswrapper[4629]: E1211 09:08:25.563267 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea77da00-5dbb-4f8a-883c-3c8984c55122" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.563290 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea77da00-5dbb-4f8a-883c-3c8984c55122" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.563538 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea77da00-5dbb-4f8a-883c-3c8984c55122" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.564532 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.567235 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.568124 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.568220 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.569578 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qjfc4" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.576977 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt"] Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.709773 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.709895 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkwzb\" (UniqueName: \"kubernetes.io/projected/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-kube-api-access-zkwzb\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.710172 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.811910 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.812313 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.812446 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkwzb\" (UniqueName: \"kubernetes.io/projected/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-kube-api-access-zkwzb\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.816129 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.816630 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.830098 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkwzb\" (UniqueName: \"kubernetes.io/projected/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-kube-api-access-zkwzb\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:25 crc kubenswrapper[4629]: I1211 09:08:25.881844 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:26 crc kubenswrapper[4629]: I1211 09:08:26.078228 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-vr26n"] Dec 11 09:08:26 crc kubenswrapper[4629]: I1211 09:08:26.098771 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-vr26n"] Dec 11 09:08:26 crc kubenswrapper[4629]: I1211 09:08:26.211735 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c6d6e77-ab1f-494b-9d78-78544b6068c4" path="/var/lib/kubelet/pods/9c6d6e77-ab1f-494b-9d78-78544b6068c4/volumes" Dec 11 09:08:26 crc kubenswrapper[4629]: I1211 09:08:26.533625 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt"] Dec 11 09:08:27 crc kubenswrapper[4629]: I1211 09:08:27.517555 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" event={"ID":"22acf22a-3fc4-4cf4-88e1-bbd5863c906c","Type":"ContainerStarted","Data":"a9cba986903d1064fb5b1f04db7a1d4f6633379bf08d776d78e669d7d53b4ec0"} Dec 11 09:08:27 crc kubenswrapper[4629]: I1211 09:08:27.517959 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" event={"ID":"22acf22a-3fc4-4cf4-88e1-bbd5863c906c","Type":"ContainerStarted","Data":"540d700bac9be13a3a9aa4b51a7b8edc7b69242f847f97a55ce7e849340e3e8f"} Dec 11 09:08:27 crc kubenswrapper[4629]: I1211 09:08:27.556051 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" podStartSLOduration=2.042440971 podStartE2EDuration="2.556025528s" podCreationTimestamp="2025-12-11 09:08:25 +0000 UTC" firstStartedPulling="2025-12-11 09:08:26.537918877 +0000 UTC m=+1894.603334485" lastFinishedPulling="2025-12-11 09:08:27.051503394 +0000 UTC m=+1895.116919042" observedRunningTime="2025-12-11 09:08:27.550611538 +0000 UTC m=+1895.616027146" watchObservedRunningTime="2025-12-11 09:08:27.556025528 +0000 UTC m=+1895.621441136" Dec 11 09:08:29 crc kubenswrapper[4629]: I1211 09:08:29.199385 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:08:29 crc kubenswrapper[4629]: E1211 09:08:29.199973 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:08:33 crc kubenswrapper[4629]: I1211 09:08:33.574907 4629 generic.go:334] "Generic (PLEG): container finished" podID="22acf22a-3fc4-4cf4-88e1-bbd5863c906c" containerID="a9cba986903d1064fb5b1f04db7a1d4f6633379bf08d776d78e669d7d53b4ec0" exitCode=0 Dec 11 09:08:33 crc kubenswrapper[4629]: I1211 09:08:33.575022 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" event={"ID":"22acf22a-3fc4-4cf4-88e1-bbd5863c906c","Type":"ContainerDied","Data":"a9cba986903d1064fb5b1f04db7a1d4f6633379bf08d776d78e669d7d53b4ec0"} Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.079527 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.271653 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkwzb\" (UniqueName: \"kubernetes.io/projected/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-kube-api-access-zkwzb\") pod \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.271793 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-ssh-key\") pod \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.271869 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-inventory\") pod \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.283817 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-kube-api-access-zkwzb" (OuterVolumeSpecName: "kube-api-access-zkwzb") pod "22acf22a-3fc4-4cf4-88e1-bbd5863c906c" (UID: "22acf22a-3fc4-4cf4-88e1-bbd5863c906c"). InnerVolumeSpecName "kube-api-access-zkwzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:08:35 crc kubenswrapper[4629]: E1211 09:08:35.299828 4629 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-inventory podName:22acf22a-3fc4-4cf4-88e1-bbd5863c906c nodeName:}" failed. No retries permitted until 2025-12-11 09:08:35.799765963 +0000 UTC m=+1903.865181571 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-inventory") pod "22acf22a-3fc4-4cf4-88e1-bbd5863c906c" (UID: "22acf22a-3fc4-4cf4-88e1-bbd5863c906c") : error deleting /var/lib/kubelet/pods/22acf22a-3fc4-4cf4-88e1-bbd5863c906c/volume-subpaths: remove /var/lib/kubelet/pods/22acf22a-3fc4-4cf4-88e1-bbd5863c906c/volume-subpaths: no such file or directory Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.302703 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "22acf22a-3fc4-4cf4-88e1-bbd5863c906c" (UID: "22acf22a-3fc4-4cf4-88e1-bbd5863c906c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.373888 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkwzb\" (UniqueName: \"kubernetes.io/projected/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-kube-api-access-zkwzb\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.374264 4629 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.593431 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" event={"ID":"22acf22a-3fc4-4cf4-88e1-bbd5863c906c","Type":"ContainerDied","Data":"540d700bac9be13a3a9aa4b51a7b8edc7b69242f847f97a55ce7e849340e3e8f"} Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.593482 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.593506 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="540d700bac9be13a3a9aa4b51a7b8edc7b69242f847f97a55ce7e849340e3e8f" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.672955 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk"] Dec 11 09:08:35 crc kubenswrapper[4629]: E1211 09:08:35.673421 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22acf22a-3fc4-4cf4-88e1-bbd5863c906c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.673443 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="22acf22a-3fc4-4cf4-88e1-bbd5863c906c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.673689 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="22acf22a-3fc4-4cf4-88e1-bbd5863c906c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.674530 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.735592 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk"] Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.780368 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f417ce11-881e-4571-913b-3b41106c935e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xrltk\" (UID: \"f417ce11-881e-4571-913b-3b41106c935e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.780450 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pd79\" (UniqueName: \"kubernetes.io/projected/f417ce11-881e-4571-913b-3b41106c935e-kube-api-access-7pd79\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xrltk\" (UID: \"f417ce11-881e-4571-913b-3b41106c935e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.780514 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f417ce11-881e-4571-913b-3b41106c935e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xrltk\" (UID: \"f417ce11-881e-4571-913b-3b41106c935e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.881972 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-inventory\") pod \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\" (UID: \"22acf22a-3fc4-4cf4-88e1-bbd5863c906c\") " Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.882755 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f417ce11-881e-4571-913b-3b41106c935e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xrltk\" (UID: \"f417ce11-881e-4571-913b-3b41106c935e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.883043 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f417ce11-881e-4571-913b-3b41106c935e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xrltk\" (UID: \"f417ce11-881e-4571-913b-3b41106c935e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.883194 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pd79\" (UniqueName: \"kubernetes.io/projected/f417ce11-881e-4571-913b-3b41106c935e-kube-api-access-7pd79\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xrltk\" (UID: \"f417ce11-881e-4571-913b-3b41106c935e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.886651 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-inventory" (OuterVolumeSpecName: "inventory") pod "22acf22a-3fc4-4cf4-88e1-bbd5863c906c" (UID: "22acf22a-3fc4-4cf4-88e1-bbd5863c906c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.887381 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f417ce11-881e-4571-913b-3b41106c935e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xrltk\" (UID: \"f417ce11-881e-4571-913b-3b41106c935e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.893276 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f417ce11-881e-4571-913b-3b41106c935e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xrltk\" (UID: \"f417ce11-881e-4571-913b-3b41106c935e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.902834 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pd79\" (UniqueName: \"kubernetes.io/projected/f417ce11-881e-4571-913b-3b41106c935e-kube-api-access-7pd79\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xrltk\" (UID: \"f417ce11-881e-4571-913b-3b41106c935e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:08:35 crc kubenswrapper[4629]: I1211 09:08:35.984239 4629 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22acf22a-3fc4-4cf4-88e1-bbd5863c906c-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:36 crc kubenswrapper[4629]: I1211 09:08:36.004398 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:08:36 crc kubenswrapper[4629]: I1211 09:08:36.539541 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk"] Dec 11 09:08:36 crc kubenswrapper[4629]: I1211 09:08:36.603099 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" event={"ID":"f417ce11-881e-4571-913b-3b41106c935e","Type":"ContainerStarted","Data":"62ddcec6d98998e0e72194ef65c8e0b4357bbb6c705bc3375cd25e9143373d1d"} Dec 11 09:08:37 crc kubenswrapper[4629]: I1211 09:08:37.613930 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" event={"ID":"f417ce11-881e-4571-913b-3b41106c935e","Type":"ContainerStarted","Data":"1416f0a9fb03c893587e5eed82173258590074ab2dcb9c6321c0924f396b5ed8"} Dec 11 09:08:37 crc kubenswrapper[4629]: I1211 09:08:37.633535 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" podStartSLOduration=2.138244476 podStartE2EDuration="2.633511239s" podCreationTimestamp="2025-12-11 09:08:35 +0000 UTC" firstStartedPulling="2025-12-11 09:08:36.551521661 +0000 UTC m=+1904.616937269" lastFinishedPulling="2025-12-11 09:08:37.046788424 +0000 UTC m=+1905.112204032" observedRunningTime="2025-12-11 09:08:37.633433136 +0000 UTC m=+1905.698848754" watchObservedRunningTime="2025-12-11 09:08:37.633511239 +0000 UTC m=+1905.698926847" Dec 11 09:08:44 crc kubenswrapper[4629]: I1211 09:08:44.199540 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:08:44 crc kubenswrapper[4629]: E1211 09:08:44.200551 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:08:58 crc kubenswrapper[4629]: I1211 09:08:58.198880 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:08:58 crc kubenswrapper[4629]: I1211 09:08:58.880383 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"3070d82dc8c3e310ce3a87a010814089bcbb734e3b66f41c4a60ff1e08bb1f40"} Dec 11 09:09:05 crc kubenswrapper[4629]: I1211 09:09:05.052949 4629 scope.go:117] "RemoveContainer" containerID="ad1c3c3c6381b9c31e7e33dd0b2fc76c425743692ff56017681087e572bfcbee" Dec 11 09:09:05 crc kubenswrapper[4629]: I1211 09:09:05.080274 4629 scope.go:117] "RemoveContainer" containerID="7922a10930f04c05077ed31db686e72349c237483cbd8679b00a579b68b02711" Dec 11 09:09:05 crc kubenswrapper[4629]: I1211 09:09:05.117359 4629 scope.go:117] "RemoveContainer" containerID="a4071f778afdc5a4701bc2dde7495a9a86f86574a817bca36c382710315f93f8" Dec 11 09:09:05 crc kubenswrapper[4629]: I1211 09:09:05.163004 4629 scope.go:117] "RemoveContainer" containerID="85efd81d390c7a8ea17a19e461389047e42fbec13324724a785a1edcfb045300" Dec 11 09:09:05 crc kubenswrapper[4629]: I1211 09:09:05.191231 4629 scope.go:117] "RemoveContainer" containerID="4a234fb7b276826369e032cc2731d56fbbc927b19e0a4e6f85f00953f805f809" Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.068837 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-frlzd"] Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.081114 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-lvfl4"] Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.089424 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-8a9c-account-create-update-f8jwt"] Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.097894 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-lvfl4"] Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.106511 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-8a9c-account-create-update-f8jwt"] Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.115229 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-6448-account-create-update-jgskv"] Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.125976 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-7a56-account-create-update-kxj5q"] Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.137226 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-2s8wp"] Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.149009 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-frlzd"] Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.158464 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-7a56-account-create-update-kxj5q"] Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.170335 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-2s8wp"] Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.180794 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-6448-account-create-update-jgskv"] Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.211079 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="373ffecc-c412-4c1c-9451-6c19a39cd8b3" path="/var/lib/kubelet/pods/373ffecc-c412-4c1c-9451-6c19a39cd8b3/volumes" Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.212225 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40733a88-03b2-45f0-b982-d6a00fefc036" path="/var/lib/kubelet/pods/40733a88-03b2-45f0-b982-d6a00fefc036/volumes" Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.213086 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fe1ea45-117b-45a7-bae8-2ef0f78660f2" path="/var/lib/kubelet/pods/6fe1ea45-117b-45a7-bae8-2ef0f78660f2/volumes" Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.213924 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f738746-49e3-4f84-a85d-124100f4f9de" path="/var/lib/kubelet/pods/8f738746-49e3-4f84-a85d-124100f4f9de/volumes" Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.215452 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f784d8f-ce06-4666-8786-02f327054c68" path="/var/lib/kubelet/pods/8f784d8f-ce06-4666-8786-02f327054c68/volumes" Dec 11 09:09:16 crc kubenswrapper[4629]: I1211 09:09:16.216505 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcdb2ea1-050a-4541-af4d-db3f2166f578" path="/var/lib/kubelet/pods/bcdb2ea1-050a-4541-af4d-db3f2166f578/volumes" Dec 11 09:09:21 crc kubenswrapper[4629]: I1211 09:09:21.169490 4629 generic.go:334] "Generic (PLEG): container finished" podID="f417ce11-881e-4571-913b-3b41106c935e" containerID="1416f0a9fb03c893587e5eed82173258590074ab2dcb9c6321c0924f396b5ed8" exitCode=0 Dec 11 09:09:21 crc kubenswrapper[4629]: I1211 09:09:21.169636 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" event={"ID":"f417ce11-881e-4571-913b-3b41106c935e","Type":"ContainerDied","Data":"1416f0a9fb03c893587e5eed82173258590074ab2dcb9c6321c0924f396b5ed8"} Dec 11 09:09:22 crc kubenswrapper[4629]: I1211 09:09:22.627963 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:09:22 crc kubenswrapper[4629]: I1211 09:09:22.919981 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pd79\" (UniqueName: \"kubernetes.io/projected/f417ce11-881e-4571-913b-3b41106c935e-kube-api-access-7pd79\") pod \"f417ce11-881e-4571-913b-3b41106c935e\" (UID: \"f417ce11-881e-4571-913b-3b41106c935e\") " Dec 11 09:09:22 crc kubenswrapper[4629]: I1211 09:09:22.920148 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f417ce11-881e-4571-913b-3b41106c935e-ssh-key\") pod \"f417ce11-881e-4571-913b-3b41106c935e\" (UID: \"f417ce11-881e-4571-913b-3b41106c935e\") " Dec 11 09:09:22 crc kubenswrapper[4629]: I1211 09:09:22.920237 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f417ce11-881e-4571-913b-3b41106c935e-inventory\") pod \"f417ce11-881e-4571-913b-3b41106c935e\" (UID: \"f417ce11-881e-4571-913b-3b41106c935e\") " Dec 11 09:09:22 crc kubenswrapper[4629]: I1211 09:09:22.929349 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f417ce11-881e-4571-913b-3b41106c935e-kube-api-access-7pd79" (OuterVolumeSpecName: "kube-api-access-7pd79") pod "f417ce11-881e-4571-913b-3b41106c935e" (UID: "f417ce11-881e-4571-913b-3b41106c935e"). InnerVolumeSpecName "kube-api-access-7pd79". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:09:22 crc kubenswrapper[4629]: I1211 09:09:22.958056 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f417ce11-881e-4571-913b-3b41106c935e-inventory" (OuterVolumeSpecName: "inventory") pod "f417ce11-881e-4571-913b-3b41106c935e" (UID: "f417ce11-881e-4571-913b-3b41106c935e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:09:22 crc kubenswrapper[4629]: I1211 09:09:22.970179 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f417ce11-881e-4571-913b-3b41106c935e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f417ce11-881e-4571-913b-3b41106c935e" (UID: "f417ce11-881e-4571-913b-3b41106c935e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.023154 4629 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f417ce11-881e-4571-913b-3b41106c935e-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.023226 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pd79\" (UniqueName: \"kubernetes.io/projected/f417ce11-881e-4571-913b-3b41106c935e-kube-api-access-7pd79\") on node \"crc\" DevicePath \"\"" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.023280 4629 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f417ce11-881e-4571-913b-3b41106c935e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.190513 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.190510 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xrltk" event={"ID":"f417ce11-881e-4571-913b-3b41106c935e","Type":"ContainerDied","Data":"62ddcec6d98998e0e72194ef65c8e0b4357bbb6c705bc3375cd25e9143373d1d"} Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.191160 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62ddcec6d98998e0e72194ef65c8e0b4357bbb6c705bc3375cd25e9143373d1d" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.295213 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng"] Dec 11 09:09:23 crc kubenswrapper[4629]: E1211 09:09:23.295631 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f417ce11-881e-4571-913b-3b41106c935e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.295650 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f417ce11-881e-4571-913b-3b41106c935e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.295872 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f417ce11-881e-4571-913b-3b41106c935e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.296562 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.298752 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.298906 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qjfc4" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.302757 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.303160 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.352354 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng"] Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.435303 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5692012-6ddd-4436-8f60-bb0ebba59d81-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng\" (UID: \"c5692012-6ddd-4436-8f60-bb0ebba59d81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.435380 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5692012-6ddd-4436-8f60-bb0ebba59d81-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng\" (UID: \"c5692012-6ddd-4436-8f60-bb0ebba59d81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.435492 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4sx7\" (UniqueName: \"kubernetes.io/projected/c5692012-6ddd-4436-8f60-bb0ebba59d81-kube-api-access-n4sx7\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng\" (UID: \"c5692012-6ddd-4436-8f60-bb0ebba59d81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.571377 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5692012-6ddd-4436-8f60-bb0ebba59d81-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng\" (UID: \"c5692012-6ddd-4436-8f60-bb0ebba59d81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.571509 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5692012-6ddd-4436-8f60-bb0ebba59d81-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng\" (UID: \"c5692012-6ddd-4436-8f60-bb0ebba59d81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.571657 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4sx7\" (UniqueName: \"kubernetes.io/projected/c5692012-6ddd-4436-8f60-bb0ebba59d81-kube-api-access-n4sx7\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng\" (UID: \"c5692012-6ddd-4436-8f60-bb0ebba59d81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.579152 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5692012-6ddd-4436-8f60-bb0ebba59d81-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng\" (UID: \"c5692012-6ddd-4436-8f60-bb0ebba59d81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.579712 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5692012-6ddd-4436-8f60-bb0ebba59d81-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng\" (UID: \"c5692012-6ddd-4436-8f60-bb0ebba59d81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.604521 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4sx7\" (UniqueName: \"kubernetes.io/projected/c5692012-6ddd-4436-8f60-bb0ebba59d81-kube-api-access-n4sx7\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng\" (UID: \"c5692012-6ddd-4436-8f60-bb0ebba59d81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:23 crc kubenswrapper[4629]: I1211 09:09:23.620440 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:24 crc kubenswrapper[4629]: I1211 09:09:24.224234 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng"] Dec 11 09:09:25 crc kubenswrapper[4629]: I1211 09:09:25.214216 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" event={"ID":"c5692012-6ddd-4436-8f60-bb0ebba59d81","Type":"ContainerStarted","Data":"4076589f3ffbac3815ae8a4b2241b9be0b6171122039d3bad024168eb598c798"} Dec 11 09:09:26 crc kubenswrapper[4629]: I1211 09:09:26.224041 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" event={"ID":"c5692012-6ddd-4436-8f60-bb0ebba59d81","Type":"ContainerStarted","Data":"82c494acc2013e6e72f8628fc5f56ad3d2e5d73adc93434d57ed7c1c2b925d1e"} Dec 11 09:09:26 crc kubenswrapper[4629]: I1211 09:09:26.240841 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" podStartSLOduration=2.362209513 podStartE2EDuration="3.240817234s" podCreationTimestamp="2025-12-11 09:09:23 +0000 UTC" firstStartedPulling="2025-12-11 09:09:24.232404254 +0000 UTC m=+1952.297819862" lastFinishedPulling="2025-12-11 09:09:25.111011975 +0000 UTC m=+1953.176427583" observedRunningTime="2025-12-11 09:09:26.237731207 +0000 UTC m=+1954.303146825" watchObservedRunningTime="2025-12-11 09:09:26.240817234 +0000 UTC m=+1954.306232842" Dec 11 09:09:30 crc kubenswrapper[4629]: I1211 09:09:30.258433 4629 generic.go:334] "Generic (PLEG): container finished" podID="c5692012-6ddd-4436-8f60-bb0ebba59d81" containerID="82c494acc2013e6e72f8628fc5f56ad3d2e5d73adc93434d57ed7c1c2b925d1e" exitCode=0 Dec 11 09:09:30 crc kubenswrapper[4629]: I1211 09:09:30.258523 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" event={"ID":"c5692012-6ddd-4436-8f60-bb0ebba59d81","Type":"ContainerDied","Data":"82c494acc2013e6e72f8628fc5f56ad3d2e5d73adc93434d57ed7c1c2b925d1e"} Dec 11 09:09:31 crc kubenswrapper[4629]: I1211 09:09:31.682226 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:31 crc kubenswrapper[4629]: I1211 09:09:31.859367 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5692012-6ddd-4436-8f60-bb0ebba59d81-inventory\") pod \"c5692012-6ddd-4436-8f60-bb0ebba59d81\" (UID: \"c5692012-6ddd-4436-8f60-bb0ebba59d81\") " Dec 11 09:09:31 crc kubenswrapper[4629]: I1211 09:09:31.860092 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5692012-6ddd-4436-8f60-bb0ebba59d81-ssh-key\") pod \"c5692012-6ddd-4436-8f60-bb0ebba59d81\" (UID: \"c5692012-6ddd-4436-8f60-bb0ebba59d81\") " Dec 11 09:09:31 crc kubenswrapper[4629]: I1211 09:09:31.860150 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4sx7\" (UniqueName: \"kubernetes.io/projected/c5692012-6ddd-4436-8f60-bb0ebba59d81-kube-api-access-n4sx7\") pod \"c5692012-6ddd-4436-8f60-bb0ebba59d81\" (UID: \"c5692012-6ddd-4436-8f60-bb0ebba59d81\") " Dec 11 09:09:31 crc kubenswrapper[4629]: I1211 09:09:31.865541 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5692012-6ddd-4436-8f60-bb0ebba59d81-kube-api-access-n4sx7" (OuterVolumeSpecName: "kube-api-access-n4sx7") pod "c5692012-6ddd-4436-8f60-bb0ebba59d81" (UID: "c5692012-6ddd-4436-8f60-bb0ebba59d81"). InnerVolumeSpecName "kube-api-access-n4sx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:09:31 crc kubenswrapper[4629]: I1211 09:09:31.894415 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5692012-6ddd-4436-8f60-bb0ebba59d81-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c5692012-6ddd-4436-8f60-bb0ebba59d81" (UID: "c5692012-6ddd-4436-8f60-bb0ebba59d81"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:09:31 crc kubenswrapper[4629]: I1211 09:09:31.900434 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5692012-6ddd-4436-8f60-bb0ebba59d81-inventory" (OuterVolumeSpecName: "inventory") pod "c5692012-6ddd-4436-8f60-bb0ebba59d81" (UID: "c5692012-6ddd-4436-8f60-bb0ebba59d81"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:09:31 crc kubenswrapper[4629]: I1211 09:09:31.962570 4629 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5692012-6ddd-4436-8f60-bb0ebba59d81-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 09:09:31 crc kubenswrapper[4629]: I1211 09:09:31.962607 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4sx7\" (UniqueName: \"kubernetes.io/projected/c5692012-6ddd-4436-8f60-bb0ebba59d81-kube-api-access-n4sx7\") on node \"crc\" DevicePath \"\"" Dec 11 09:09:31 crc kubenswrapper[4629]: I1211 09:09:31.962620 4629 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5692012-6ddd-4436-8f60-bb0ebba59d81-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.274737 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" event={"ID":"c5692012-6ddd-4436-8f60-bb0ebba59d81","Type":"ContainerDied","Data":"4076589f3ffbac3815ae8a4b2241b9be0b6171122039d3bad024168eb598c798"} Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.274794 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4076589f3ffbac3815ae8a4b2241b9be0b6171122039d3bad024168eb598c798" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.274800 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.366715 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld"] Dec 11 09:09:32 crc kubenswrapper[4629]: E1211 09:09:32.367401 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5692012-6ddd-4436-8f60-bb0ebba59d81" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.367433 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5692012-6ddd-4436-8f60-bb0ebba59d81" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.367679 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5692012-6ddd-4436-8f60-bb0ebba59d81" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.368508 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.372180 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.374531 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.374897 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.379965 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qjfc4" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.386207 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld"] Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.470223 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7gkld\" (UID: \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.470299 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj9cw\" (UniqueName: \"kubernetes.io/projected/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-kube-api-access-jj9cw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7gkld\" (UID: \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.470377 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7gkld\" (UID: \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.572401 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7gkld\" (UID: \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.572575 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7gkld\" (UID: \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.572618 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj9cw\" (UniqueName: \"kubernetes.io/projected/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-kube-api-access-jj9cw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7gkld\" (UID: \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.580780 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7gkld\" (UID: \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.581648 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7gkld\" (UID: \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.590778 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj9cw\" (UniqueName: \"kubernetes.io/projected/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-kube-api-access-jj9cw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7gkld\" (UID: \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:09:32 crc kubenswrapper[4629]: I1211 09:09:32.712575 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:09:33 crc kubenswrapper[4629]: I1211 09:09:33.496373 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld"] Dec 11 09:09:34 crc kubenswrapper[4629]: I1211 09:09:34.294111 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" event={"ID":"99f4ba2c-fa6c-438f-8c01-90663ffc9f80","Type":"ContainerStarted","Data":"96ac85d806faec69fe96a014460801b865ea0b804ca799e95c4d9bf3eac327c5"} Dec 11 09:09:35 crc kubenswrapper[4629]: I1211 09:09:35.303799 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" event={"ID":"99f4ba2c-fa6c-438f-8c01-90663ffc9f80","Type":"ContainerStarted","Data":"587dc45d34e049d9f19b4e1f50b928e34e7d2f22f6aff1e75698c9896ece746c"} Dec 11 09:09:35 crc kubenswrapper[4629]: I1211 09:09:35.327013 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" podStartSLOduration=2.616574026 podStartE2EDuration="3.326991625s" podCreationTimestamp="2025-12-11 09:09:32 +0000 UTC" firstStartedPulling="2025-12-11 09:09:33.484141425 +0000 UTC m=+1961.549557033" lastFinishedPulling="2025-12-11 09:09:34.194559024 +0000 UTC m=+1962.259974632" observedRunningTime="2025-12-11 09:09:35.321227295 +0000 UTC m=+1963.386642903" watchObservedRunningTime="2025-12-11 09:09:35.326991625 +0000 UTC m=+1963.392407233" Dec 11 09:09:52 crc kubenswrapper[4629]: I1211 09:09:52.065766 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g6gr5"] Dec 11 09:09:52 crc kubenswrapper[4629]: I1211 09:09:52.076480 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g6gr5"] Dec 11 09:09:52 crc kubenswrapper[4629]: I1211 09:09:52.212893 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45f8baeb-687e-4c40-8ea9-05e66a26c9a7" path="/var/lib/kubelet/pods/45f8baeb-687e-4c40-8ea9-05e66a26c9a7/volumes" Dec 11 09:10:05 crc kubenswrapper[4629]: I1211 09:10:05.324460 4629 scope.go:117] "RemoveContainer" containerID="8698a7dcb6d1bf64f0e490c80eb1f24ee35ba2acee78dc41795884c93867b8b1" Dec 11 09:10:05 crc kubenswrapper[4629]: I1211 09:10:05.355316 4629 scope.go:117] "RemoveContainer" containerID="3d6b02d32cae382d0020b0b30f12e0fddbfabd7aaf0cebb386145524866a6ad7" Dec 11 09:10:05 crc kubenswrapper[4629]: I1211 09:10:05.402738 4629 scope.go:117] "RemoveContainer" containerID="828f2f621c907f59d2698842ac2e387d0a2be413afe971e677dde5f123ae7eca" Dec 11 09:10:05 crc kubenswrapper[4629]: I1211 09:10:05.451553 4629 scope.go:117] "RemoveContainer" containerID="91ffd72b0961ed31745502586d4d8ddfc25313b7cf1442f3495c7b21f64208c5" Dec 11 09:10:05 crc kubenswrapper[4629]: I1211 09:10:05.535915 4629 scope.go:117] "RemoveContainer" containerID="fe76614364efaac6558506d4ef88813004df0129c350fd46648ae5610a1205fe" Dec 11 09:10:05 crc kubenswrapper[4629]: I1211 09:10:05.561207 4629 scope.go:117] "RemoveContainer" containerID="6cecbef414779753a6e47aa24ccfed626aecffadf7dfa358b2f274ffa2d77a10" Dec 11 09:10:05 crc kubenswrapper[4629]: I1211 09:10:05.603672 4629 scope.go:117] "RemoveContainer" containerID="fd100bb683f86cee57c19258890d16d1c0566b0c2ac3edc7b27b138f76273aa3" Dec 11 09:10:19 crc kubenswrapper[4629]: I1211 09:10:19.055044 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-vppnj"] Dec 11 09:10:19 crc kubenswrapper[4629]: I1211 09:10:19.060344 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-vppnj"] Dec 11 09:10:20 crc kubenswrapper[4629]: I1211 09:10:20.028935 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-65bhx"] Dec 11 09:10:20 crc kubenswrapper[4629]: I1211 09:10:20.036081 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-65bhx"] Dec 11 09:10:20 crc kubenswrapper[4629]: I1211 09:10:20.209858 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f83bf9d-806c-4310-9315-2bcbd529a6d1" path="/var/lib/kubelet/pods/8f83bf9d-806c-4310-9315-2bcbd529a6d1/volumes" Dec 11 09:10:20 crc kubenswrapper[4629]: I1211 09:10:20.210942 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5f5f734-35d7-49f9-a007-d1722d4d4673" path="/var/lib/kubelet/pods/a5f5f734-35d7-49f9-a007-d1722d4d4673/volumes" Dec 11 09:10:28 crc kubenswrapper[4629]: I1211 09:10:28.754584 4629 generic.go:334] "Generic (PLEG): container finished" podID="99f4ba2c-fa6c-438f-8c01-90663ffc9f80" containerID="587dc45d34e049d9f19b4e1f50b928e34e7d2f22f6aff1e75698c9896ece746c" exitCode=0 Dec 11 09:10:28 crc kubenswrapper[4629]: I1211 09:10:28.754683 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" event={"ID":"99f4ba2c-fa6c-438f-8c01-90663ffc9f80","Type":"ContainerDied","Data":"587dc45d34e049d9f19b4e1f50b928e34e7d2f22f6aff1e75698c9896ece746c"} Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.241221 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.426237 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-ssh-key\") pod \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\" (UID: \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\") " Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.426293 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj9cw\" (UniqueName: \"kubernetes.io/projected/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-kube-api-access-jj9cw\") pod \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\" (UID: \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\") " Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.426519 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-inventory\") pod \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\" (UID: \"99f4ba2c-fa6c-438f-8c01-90663ffc9f80\") " Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.445315 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-kube-api-access-jj9cw" (OuterVolumeSpecName: "kube-api-access-jj9cw") pod "99f4ba2c-fa6c-438f-8c01-90663ffc9f80" (UID: "99f4ba2c-fa6c-438f-8c01-90663ffc9f80"). InnerVolumeSpecName "kube-api-access-jj9cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.458987 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-inventory" (OuterVolumeSpecName: "inventory") pod "99f4ba2c-fa6c-438f-8c01-90663ffc9f80" (UID: "99f4ba2c-fa6c-438f-8c01-90663ffc9f80"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.489372 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "99f4ba2c-fa6c-438f-8c01-90663ffc9f80" (UID: "99f4ba2c-fa6c-438f-8c01-90663ffc9f80"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.528714 4629 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.528753 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj9cw\" (UniqueName: \"kubernetes.io/projected/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-kube-api-access-jj9cw\") on node \"crc\" DevicePath \"\"" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.528767 4629 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99f4ba2c-fa6c-438f-8c01-90663ffc9f80-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.774568 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" event={"ID":"99f4ba2c-fa6c-438f-8c01-90663ffc9f80","Type":"ContainerDied","Data":"96ac85d806faec69fe96a014460801b865ea0b804ca799e95c4d9bf3eac327c5"} Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.774634 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96ac85d806faec69fe96a014460801b865ea0b804ca799e95c4d9bf3eac327c5" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.774676 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7gkld" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.871478 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ztllp"] Dec 11 09:10:30 crc kubenswrapper[4629]: E1211 09:10:30.872004 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99f4ba2c-fa6c-438f-8c01-90663ffc9f80" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.872032 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="99f4ba2c-fa6c-438f-8c01-90663ffc9f80" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.872283 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="99f4ba2c-fa6c-438f-8c01-90663ffc9f80" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.873378 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.875775 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.875894 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.875918 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.876724 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qjfc4" Dec 11 09:10:30 crc kubenswrapper[4629]: I1211 09:10:30.884008 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ztllp"] Dec 11 09:10:31 crc kubenswrapper[4629]: I1211 09:10:31.037297 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c5239464-94d3-4812-aad7-6c864eb46731-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ztllp\" (UID: \"c5239464-94d3-4812-aad7-6c864eb46731\") " pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:31 crc kubenswrapper[4629]: I1211 09:10:31.037374 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c5239464-94d3-4812-aad7-6c864eb46731-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ztllp\" (UID: \"c5239464-94d3-4812-aad7-6c864eb46731\") " pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:31 crc kubenswrapper[4629]: I1211 09:10:31.037510 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgjg2\" (UniqueName: \"kubernetes.io/projected/c5239464-94d3-4812-aad7-6c864eb46731-kube-api-access-cgjg2\") pod \"ssh-known-hosts-edpm-deployment-ztllp\" (UID: \"c5239464-94d3-4812-aad7-6c864eb46731\") " pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:31 crc kubenswrapper[4629]: I1211 09:10:31.139132 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c5239464-94d3-4812-aad7-6c864eb46731-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ztllp\" (UID: \"c5239464-94d3-4812-aad7-6c864eb46731\") " pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:31 crc kubenswrapper[4629]: I1211 09:10:31.139206 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c5239464-94d3-4812-aad7-6c864eb46731-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ztllp\" (UID: \"c5239464-94d3-4812-aad7-6c864eb46731\") " pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:31 crc kubenswrapper[4629]: I1211 09:10:31.139288 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgjg2\" (UniqueName: \"kubernetes.io/projected/c5239464-94d3-4812-aad7-6c864eb46731-kube-api-access-cgjg2\") pod \"ssh-known-hosts-edpm-deployment-ztllp\" (UID: \"c5239464-94d3-4812-aad7-6c864eb46731\") " pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:31 crc kubenswrapper[4629]: I1211 09:10:31.144082 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c5239464-94d3-4812-aad7-6c864eb46731-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ztllp\" (UID: \"c5239464-94d3-4812-aad7-6c864eb46731\") " pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:31 crc kubenswrapper[4629]: I1211 09:10:31.157822 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgjg2\" (UniqueName: \"kubernetes.io/projected/c5239464-94d3-4812-aad7-6c864eb46731-kube-api-access-cgjg2\") pod \"ssh-known-hosts-edpm-deployment-ztllp\" (UID: \"c5239464-94d3-4812-aad7-6c864eb46731\") " pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:31 crc kubenswrapper[4629]: I1211 09:10:31.161987 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c5239464-94d3-4812-aad7-6c864eb46731-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ztllp\" (UID: \"c5239464-94d3-4812-aad7-6c864eb46731\") " pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:31 crc kubenswrapper[4629]: I1211 09:10:31.196584 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:31 crc kubenswrapper[4629]: I1211 09:10:31.705683 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ztllp"] Dec 11 09:10:31 crc kubenswrapper[4629]: I1211 09:10:31.784065 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" event={"ID":"c5239464-94d3-4812-aad7-6c864eb46731","Type":"ContainerStarted","Data":"8493b2b7d024afb279c0c18876628b63ae28985fb92258862e035c31d66171ab"} Dec 11 09:10:32 crc kubenswrapper[4629]: I1211 09:10:32.795825 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" event={"ID":"c5239464-94d3-4812-aad7-6c864eb46731","Type":"ContainerStarted","Data":"766d890525d37955a3666dd481b29c9f160c2ef50a1aef740d7251606dc0f7a7"} Dec 11 09:10:32 crc kubenswrapper[4629]: I1211 09:10:32.824144 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" podStartSLOduration=2.27066957 podStartE2EDuration="2.82411385s" podCreationTimestamp="2025-12-11 09:10:30 +0000 UTC" firstStartedPulling="2025-12-11 09:10:31.715982901 +0000 UTC m=+2019.781398509" lastFinishedPulling="2025-12-11 09:10:32.269427181 +0000 UTC m=+2020.334842789" observedRunningTime="2025-12-11 09:10:32.818491563 +0000 UTC m=+2020.883907171" watchObservedRunningTime="2025-12-11 09:10:32.82411385 +0000 UTC m=+2020.889529458" Dec 11 09:10:40 crc kubenswrapper[4629]: I1211 09:10:40.862500 4629 generic.go:334] "Generic (PLEG): container finished" podID="c5239464-94d3-4812-aad7-6c864eb46731" containerID="766d890525d37955a3666dd481b29c9f160c2ef50a1aef740d7251606dc0f7a7" exitCode=0 Dec 11 09:10:40 crc kubenswrapper[4629]: I1211 09:10:40.863155 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" event={"ID":"c5239464-94d3-4812-aad7-6c864eb46731","Type":"ContainerDied","Data":"766d890525d37955a3666dd481b29c9f160c2ef50a1aef740d7251606dc0f7a7"} Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.275554 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.451370 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c5239464-94d3-4812-aad7-6c864eb46731-ssh-key-openstack-edpm-ipam\") pod \"c5239464-94d3-4812-aad7-6c864eb46731\" (UID: \"c5239464-94d3-4812-aad7-6c864eb46731\") " Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.451481 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c5239464-94d3-4812-aad7-6c864eb46731-inventory-0\") pod \"c5239464-94d3-4812-aad7-6c864eb46731\" (UID: \"c5239464-94d3-4812-aad7-6c864eb46731\") " Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.451580 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgjg2\" (UniqueName: \"kubernetes.io/projected/c5239464-94d3-4812-aad7-6c864eb46731-kube-api-access-cgjg2\") pod \"c5239464-94d3-4812-aad7-6c864eb46731\" (UID: \"c5239464-94d3-4812-aad7-6c864eb46731\") " Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.472490 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5239464-94d3-4812-aad7-6c864eb46731-kube-api-access-cgjg2" (OuterVolumeSpecName: "kube-api-access-cgjg2") pod "c5239464-94d3-4812-aad7-6c864eb46731" (UID: "c5239464-94d3-4812-aad7-6c864eb46731"). InnerVolumeSpecName "kube-api-access-cgjg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.489159 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5239464-94d3-4812-aad7-6c864eb46731-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "c5239464-94d3-4812-aad7-6c864eb46731" (UID: "c5239464-94d3-4812-aad7-6c864eb46731"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.489671 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5239464-94d3-4812-aad7-6c864eb46731-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c5239464-94d3-4812-aad7-6c864eb46731" (UID: "c5239464-94d3-4812-aad7-6c864eb46731"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.553820 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgjg2\" (UniqueName: \"kubernetes.io/projected/c5239464-94d3-4812-aad7-6c864eb46731-kube-api-access-cgjg2\") on node \"crc\" DevicePath \"\"" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.554112 4629 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c5239464-94d3-4812-aad7-6c864eb46731-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.554248 4629 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c5239464-94d3-4812-aad7-6c864eb46731-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.882440 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" event={"ID":"c5239464-94d3-4812-aad7-6c864eb46731","Type":"ContainerDied","Data":"8493b2b7d024afb279c0c18876628b63ae28985fb92258862e035c31d66171ab"} Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.882484 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8493b2b7d024afb279c0c18876628b63ae28985fb92258862e035c31d66171ab" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.882557 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ztllp" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.957244 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2"] Dec 11 09:10:42 crc kubenswrapper[4629]: E1211 09:10:42.957694 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5239464-94d3-4812-aad7-6c864eb46731" containerName="ssh-known-hosts-edpm-deployment" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.957717 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5239464-94d3-4812-aad7-6c864eb46731" containerName="ssh-known-hosts-edpm-deployment" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.957938 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5239464-94d3-4812-aad7-6c864eb46731" containerName="ssh-known-hosts-edpm-deployment" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.958670 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.961765 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.961983 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.962168 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qjfc4" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.962202 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 09:10:42 crc kubenswrapper[4629]: I1211 09:10:42.974662 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2"] Dec 11 09:10:43 crc kubenswrapper[4629]: I1211 09:10:43.061938 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v856m\" (UniqueName: \"kubernetes.io/projected/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-kube-api-access-v856m\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lmqx2\" (UID: \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:43 crc kubenswrapper[4629]: I1211 09:10:43.061990 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lmqx2\" (UID: \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:43 crc kubenswrapper[4629]: I1211 09:10:43.062055 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lmqx2\" (UID: \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:43 crc kubenswrapper[4629]: I1211 09:10:43.163436 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lmqx2\" (UID: \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:43 crc kubenswrapper[4629]: I1211 09:10:43.163596 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v856m\" (UniqueName: \"kubernetes.io/projected/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-kube-api-access-v856m\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lmqx2\" (UID: \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:43 crc kubenswrapper[4629]: I1211 09:10:43.163628 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lmqx2\" (UID: \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:43 crc kubenswrapper[4629]: I1211 09:10:43.168322 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lmqx2\" (UID: \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:43 crc kubenswrapper[4629]: I1211 09:10:43.177478 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lmqx2\" (UID: \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:43 crc kubenswrapper[4629]: I1211 09:10:43.180614 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v856m\" (UniqueName: \"kubernetes.io/projected/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-kube-api-access-v856m\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lmqx2\" (UID: \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:43 crc kubenswrapper[4629]: I1211 09:10:43.286261 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:44 crc kubenswrapper[4629]: I1211 09:10:44.043720 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2"] Dec 11 09:10:44 crc kubenswrapper[4629]: I1211 09:10:44.899018 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" event={"ID":"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a","Type":"ContainerStarted","Data":"d70bb9fb7659c9fa6b00f99defd7f1b0ccaab11c0ae9fd84b8e087f16784e739"} Dec 11 09:10:44 crc kubenswrapper[4629]: I1211 09:10:44.899340 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" event={"ID":"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a","Type":"ContainerStarted","Data":"14b1955be36ba80f55f64ed42f80d1a33abfecfbdae2b90a225f9d78f50e5075"} Dec 11 09:10:44 crc kubenswrapper[4629]: I1211 09:10:44.922549 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" podStartSLOduration=2.2919387110000002 podStartE2EDuration="2.922526944s" podCreationTimestamp="2025-12-11 09:10:42 +0000 UTC" firstStartedPulling="2025-12-11 09:10:44.055068483 +0000 UTC m=+2032.120484091" lastFinishedPulling="2025-12-11 09:10:44.685656716 +0000 UTC m=+2032.751072324" observedRunningTime="2025-12-11 09:10:44.915486333 +0000 UTC m=+2032.980901931" watchObservedRunningTime="2025-12-11 09:10:44.922526944 +0000 UTC m=+2032.987942552" Dec 11 09:10:53 crc kubenswrapper[4629]: I1211 09:10:53.979378 4629 generic.go:334] "Generic (PLEG): container finished" podID="22bb894d-6f5b-4e9e-9cca-b2d75ee4220a" containerID="d70bb9fb7659c9fa6b00f99defd7f1b0ccaab11c0ae9fd84b8e087f16784e739" exitCode=0 Dec 11 09:10:53 crc kubenswrapper[4629]: I1211 09:10:53.979441 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" event={"ID":"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a","Type":"ContainerDied","Data":"d70bb9fb7659c9fa6b00f99defd7f1b0ccaab11c0ae9fd84b8e087f16784e739"} Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.378713 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.506353 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v856m\" (UniqueName: \"kubernetes.io/projected/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-kube-api-access-v856m\") pod \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\" (UID: \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\") " Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.506450 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-inventory\") pod \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\" (UID: \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\") " Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.506527 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-ssh-key\") pod \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\" (UID: \"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a\") " Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.528137 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-kube-api-access-v856m" (OuterVolumeSpecName: "kube-api-access-v856m") pod "22bb894d-6f5b-4e9e-9cca-b2d75ee4220a" (UID: "22bb894d-6f5b-4e9e-9cca-b2d75ee4220a"). InnerVolumeSpecName "kube-api-access-v856m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.542044 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-inventory" (OuterVolumeSpecName: "inventory") pod "22bb894d-6f5b-4e9e-9cca-b2d75ee4220a" (UID: "22bb894d-6f5b-4e9e-9cca-b2d75ee4220a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.548056 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "22bb894d-6f5b-4e9e-9cca-b2d75ee4220a" (UID: "22bb894d-6f5b-4e9e-9cca-b2d75ee4220a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.609339 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v856m\" (UniqueName: \"kubernetes.io/projected/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-kube-api-access-v856m\") on node \"crc\" DevicePath \"\"" Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.609557 4629 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.609661 4629 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22bb894d-6f5b-4e9e-9cca-b2d75ee4220a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.997439 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" event={"ID":"22bb894d-6f5b-4e9e-9cca-b2d75ee4220a","Type":"ContainerDied","Data":"14b1955be36ba80f55f64ed42f80d1a33abfecfbdae2b90a225f9d78f50e5075"} Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.997905 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14b1955be36ba80f55f64ed42f80d1a33abfecfbdae2b90a225f9d78f50e5075" Dec 11 09:10:55 crc kubenswrapper[4629]: I1211 09:10:55.997543 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lmqx2" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.086920 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn"] Dec 11 09:10:56 crc kubenswrapper[4629]: E1211 09:10:56.087385 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22bb894d-6f5b-4e9e-9cca-b2d75ee4220a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.087414 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="22bb894d-6f5b-4e9e-9cca-b2d75ee4220a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.087640 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="22bb894d-6f5b-4e9e-9cca-b2d75ee4220a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.088452 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.091365 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.091525 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qjfc4" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.092332 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.092569 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.103343 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn"] Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.224114 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn\" (UID: \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.224173 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn\" (UID: \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.224347 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svnk9\" (UniqueName: \"kubernetes.io/projected/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-kube-api-access-svnk9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn\" (UID: \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.326568 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn\" (UID: \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.326668 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn\" (UID: \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.326745 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svnk9\" (UniqueName: \"kubernetes.io/projected/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-kube-api-access-svnk9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn\" (UID: \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.332621 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn\" (UID: \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.343525 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn\" (UID: \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.352978 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svnk9\" (UniqueName: \"kubernetes.io/projected/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-kube-api-access-svnk9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn\" (UID: \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:10:56 crc kubenswrapper[4629]: I1211 09:10:56.405015 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:10:57 crc kubenswrapper[4629]: I1211 09:10:57.020376 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn"] Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.014970 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" event={"ID":"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0","Type":"ContainerStarted","Data":"457c6fcc689e543973ab7119307754618cb427ad9dd9d30e29755a521a5dc4be"} Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.016299 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" event={"ID":"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0","Type":"ContainerStarted","Data":"b2d83b5d3a8247f76a4700f72f9ce78011e0412ef231737ab75e51d3c8ffe473"} Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.043377 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" podStartSLOduration=1.525372108 podStartE2EDuration="2.043357145s" podCreationTimestamp="2025-12-11 09:10:56 +0000 UTC" firstStartedPulling="2025-12-11 09:10:57.044756496 +0000 UTC m=+2045.110172104" lastFinishedPulling="2025-12-11 09:10:57.562741533 +0000 UTC m=+2045.628157141" observedRunningTime="2025-12-11 09:10:58.034154857 +0000 UTC m=+2046.099570465" watchObservedRunningTime="2025-12-11 09:10:58.043357145 +0000 UTC m=+2046.108772743" Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.500432 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7fhd7"] Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.502737 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.510658 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7fhd7"] Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.571352 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aac7e45-016c-4fa2-b383-cb06b4b843d2-catalog-content\") pod \"redhat-operators-7fhd7\" (UID: \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\") " pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.571638 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmt6z\" (UniqueName: \"kubernetes.io/projected/4aac7e45-016c-4fa2-b383-cb06b4b843d2-kube-api-access-lmt6z\") pod \"redhat-operators-7fhd7\" (UID: \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\") " pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.571743 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aac7e45-016c-4fa2-b383-cb06b4b843d2-utilities\") pod \"redhat-operators-7fhd7\" (UID: \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\") " pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.673935 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmt6z\" (UniqueName: \"kubernetes.io/projected/4aac7e45-016c-4fa2-b383-cb06b4b843d2-kube-api-access-lmt6z\") pod \"redhat-operators-7fhd7\" (UID: \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\") " pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.674022 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aac7e45-016c-4fa2-b383-cb06b4b843d2-utilities\") pod \"redhat-operators-7fhd7\" (UID: \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\") " pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.674120 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aac7e45-016c-4fa2-b383-cb06b4b843d2-catalog-content\") pod \"redhat-operators-7fhd7\" (UID: \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\") " pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.674781 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aac7e45-016c-4fa2-b383-cb06b4b843d2-catalog-content\") pod \"redhat-operators-7fhd7\" (UID: \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\") " pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.674944 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aac7e45-016c-4fa2-b383-cb06b4b843d2-utilities\") pod \"redhat-operators-7fhd7\" (UID: \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\") " pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.693697 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmt6z\" (UniqueName: \"kubernetes.io/projected/4aac7e45-016c-4fa2-b383-cb06b4b843d2-kube-api-access-lmt6z\") pod \"redhat-operators-7fhd7\" (UID: \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\") " pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:10:58 crc kubenswrapper[4629]: I1211 09:10:58.833178 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:10:59 crc kubenswrapper[4629]: W1211 09:10:59.384347 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aac7e45_016c_4fa2_b383_cb06b4b843d2.slice/crio-0b5f9025adbe04a306941e8bce9b615a9a87ef950f16905bec9d4a73ec429d6c WatchSource:0}: Error finding container 0b5f9025adbe04a306941e8bce9b615a9a87ef950f16905bec9d4a73ec429d6c: Status 404 returned error can't find the container with id 0b5f9025adbe04a306941e8bce9b615a9a87ef950f16905bec9d4a73ec429d6c Dec 11 09:10:59 crc kubenswrapper[4629]: I1211 09:10:59.395922 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7fhd7"] Dec 11 09:11:00 crc kubenswrapper[4629]: I1211 09:11:00.036520 4629 generic.go:334] "Generic (PLEG): container finished" podID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" containerID="b6b6e6987387743d1feb4caad2ef01141c57a09bbd81e56e4fb92613a7e69dad" exitCode=0 Dec 11 09:11:00 crc kubenswrapper[4629]: I1211 09:11:00.036564 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fhd7" event={"ID":"4aac7e45-016c-4fa2-b383-cb06b4b843d2","Type":"ContainerDied","Data":"b6b6e6987387743d1feb4caad2ef01141c57a09bbd81e56e4fb92613a7e69dad"} Dec 11 09:11:00 crc kubenswrapper[4629]: I1211 09:11:00.036589 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fhd7" event={"ID":"4aac7e45-016c-4fa2-b383-cb06b4b843d2","Type":"ContainerStarted","Data":"0b5f9025adbe04a306941e8bce9b615a9a87ef950f16905bec9d4a73ec429d6c"} Dec 11 09:11:01 crc kubenswrapper[4629]: I1211 09:11:01.047655 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fhd7" event={"ID":"4aac7e45-016c-4fa2-b383-cb06b4b843d2","Type":"ContainerStarted","Data":"374e4709cc1f3df211673e30472bc0de3d0a4178749a6fc9d00eb45a255598b3"} Dec 11 09:11:03 crc kubenswrapper[4629]: I1211 09:11:03.161002 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-nmjwk"] Dec 11 09:11:03 crc kubenswrapper[4629]: I1211 09:11:03.173645 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-nmjwk"] Dec 11 09:11:04 crc kubenswrapper[4629]: I1211 09:11:04.254111 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2de77a96-4287-46cc-8004-04f26169a35d" path="/var/lib/kubelet/pods/2de77a96-4287-46cc-8004-04f26169a35d/volumes" Dec 11 09:11:05 crc kubenswrapper[4629]: I1211 09:11:05.806283 4629 scope.go:117] "RemoveContainer" containerID="9f9ffba2e0fd7c28249813b71c9e26135de39e68728e184c968d7475561fa5ba" Dec 11 09:11:05 crc kubenswrapper[4629]: I1211 09:11:05.897783 4629 scope.go:117] "RemoveContainer" containerID="a42cfdb3a956accc8e6e35bf29d2aa5681ece9adba1c92efc77474b2a3fb1f98" Dec 11 09:11:05 crc kubenswrapper[4629]: I1211 09:11:05.959893 4629 scope.go:117] "RemoveContainer" containerID="7167f4bf289d9b5cd36d9756beb18bfb61651fce14cb268b6be59134ddb91ded" Dec 11 09:11:06 crc kubenswrapper[4629]: I1211 09:11:06.159395 4629 generic.go:334] "Generic (PLEG): container finished" podID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" containerID="374e4709cc1f3df211673e30472bc0de3d0a4178749a6fc9d00eb45a255598b3" exitCode=0 Dec 11 09:11:06 crc kubenswrapper[4629]: I1211 09:11:06.159499 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fhd7" event={"ID":"4aac7e45-016c-4fa2-b383-cb06b4b843d2","Type":"ContainerDied","Data":"374e4709cc1f3df211673e30472bc0de3d0a4178749a6fc9d00eb45a255598b3"} Dec 11 09:11:07 crc kubenswrapper[4629]: I1211 09:11:07.174702 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fhd7" event={"ID":"4aac7e45-016c-4fa2-b383-cb06b4b843d2","Type":"ContainerStarted","Data":"684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44"} Dec 11 09:11:07 crc kubenswrapper[4629]: I1211 09:11:07.202242 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7fhd7" podStartSLOduration=2.437159148 podStartE2EDuration="9.202218149s" podCreationTimestamp="2025-12-11 09:10:58 +0000 UTC" firstStartedPulling="2025-12-11 09:11:00.038435807 +0000 UTC m=+2048.103851415" lastFinishedPulling="2025-12-11 09:11:06.803494818 +0000 UTC m=+2054.868910416" observedRunningTime="2025-12-11 09:11:07.196969855 +0000 UTC m=+2055.262385463" watchObservedRunningTime="2025-12-11 09:11:07.202218149 +0000 UTC m=+2055.267633757" Dec 11 09:11:08 crc kubenswrapper[4629]: I1211 09:11:08.833800 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:11:08 crc kubenswrapper[4629]: I1211 09:11:08.833871 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:11:09 crc kubenswrapper[4629]: I1211 09:11:09.884006 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7fhd7" podUID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" containerName="registry-server" probeResult="failure" output=< Dec 11 09:11:09 crc kubenswrapper[4629]: timeout: failed to connect service ":50051" within 1s Dec 11 09:11:09 crc kubenswrapper[4629]: > Dec 11 09:11:10 crc kubenswrapper[4629]: I1211 09:11:10.201688 4629 generic.go:334] "Generic (PLEG): container finished" podID="636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0" containerID="457c6fcc689e543973ab7119307754618cb427ad9dd9d30e29755a521a5dc4be" exitCode=0 Dec 11 09:11:10 crc kubenswrapper[4629]: I1211 09:11:10.208796 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" event={"ID":"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0","Type":"ContainerDied","Data":"457c6fcc689e543973ab7119307754618cb427ad9dd9d30e29755a521a5dc4be"} Dec 11 09:11:11 crc kubenswrapper[4629]: I1211 09:11:11.658710 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:11:11 crc kubenswrapper[4629]: I1211 09:11:11.804631 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svnk9\" (UniqueName: \"kubernetes.io/projected/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-kube-api-access-svnk9\") pod \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\" (UID: \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\") " Dec 11 09:11:11 crc kubenswrapper[4629]: I1211 09:11:11.805281 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-ssh-key\") pod \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\" (UID: \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\") " Dec 11 09:11:11 crc kubenswrapper[4629]: I1211 09:11:11.805461 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-inventory\") pod \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\" (UID: \"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0\") " Dec 11 09:11:11 crc kubenswrapper[4629]: I1211 09:11:11.810951 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-kube-api-access-svnk9" (OuterVolumeSpecName: "kube-api-access-svnk9") pod "636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0" (UID: "636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0"). InnerVolumeSpecName "kube-api-access-svnk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:11:11 crc kubenswrapper[4629]: I1211 09:11:11.859001 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-inventory" (OuterVolumeSpecName: "inventory") pod "636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0" (UID: "636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:11:11 crc kubenswrapper[4629]: I1211 09:11:11.889224 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0" (UID: "636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:11:11 crc kubenswrapper[4629]: I1211 09:11:11.910100 4629 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 09:11:11 crc kubenswrapper[4629]: I1211 09:11:11.910139 4629 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 09:11:11 crc kubenswrapper[4629]: I1211 09:11:11.910151 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svnk9\" (UniqueName: \"kubernetes.io/projected/636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0-kube-api-access-svnk9\") on node \"crc\" DevicePath \"\"" Dec 11 09:11:12 crc kubenswrapper[4629]: I1211 09:11:12.219432 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" event={"ID":"636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0","Type":"ContainerDied","Data":"b2d83b5d3a8247f76a4700f72f9ce78011e0412ef231737ab75e51d3c8ffe473"} Dec 11 09:11:12 crc kubenswrapper[4629]: I1211 09:11:12.219495 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2d83b5d3a8247f76a4700f72f9ce78011e0412ef231737ab75e51d3c8ffe473" Dec 11 09:11:12 crc kubenswrapper[4629]: I1211 09:11:12.219567 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn" Dec 11 09:11:18 crc kubenswrapper[4629]: I1211 09:11:18.895751 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:11:18 crc kubenswrapper[4629]: I1211 09:11:18.963774 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:11:19 crc kubenswrapper[4629]: I1211 09:11:19.134653 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7fhd7"] Dec 11 09:11:20 crc kubenswrapper[4629]: I1211 09:11:20.282680 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7fhd7" podUID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" containerName="registry-server" containerID="cri-o://684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44" gracePeriod=2 Dec 11 09:11:20 crc kubenswrapper[4629]: I1211 09:11:20.486712 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:11:20 crc kubenswrapper[4629]: I1211 09:11:20.486773 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:11:20 crc kubenswrapper[4629]: I1211 09:11:20.722540 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:11:20 crc kubenswrapper[4629]: I1211 09:11:20.843868 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmt6z\" (UniqueName: \"kubernetes.io/projected/4aac7e45-016c-4fa2-b383-cb06b4b843d2-kube-api-access-lmt6z\") pod \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\" (UID: \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\") " Dec 11 09:11:20 crc kubenswrapper[4629]: I1211 09:11:20.844020 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aac7e45-016c-4fa2-b383-cb06b4b843d2-catalog-content\") pod \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\" (UID: \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\") " Dec 11 09:11:20 crc kubenswrapper[4629]: I1211 09:11:20.844073 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aac7e45-016c-4fa2-b383-cb06b4b843d2-utilities\") pod \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\" (UID: \"4aac7e45-016c-4fa2-b383-cb06b4b843d2\") " Dec 11 09:11:20 crc kubenswrapper[4629]: I1211 09:11:20.845178 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aac7e45-016c-4fa2-b383-cb06b4b843d2-utilities" (OuterVolumeSpecName: "utilities") pod "4aac7e45-016c-4fa2-b383-cb06b4b843d2" (UID: "4aac7e45-016c-4fa2-b383-cb06b4b843d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:11:20 crc kubenswrapper[4629]: I1211 09:11:20.853723 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aac7e45-016c-4fa2-b383-cb06b4b843d2-kube-api-access-lmt6z" (OuterVolumeSpecName: "kube-api-access-lmt6z") pod "4aac7e45-016c-4fa2-b383-cb06b4b843d2" (UID: "4aac7e45-016c-4fa2-b383-cb06b4b843d2"). InnerVolumeSpecName "kube-api-access-lmt6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:11:20 crc kubenswrapper[4629]: I1211 09:11:20.946723 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmt6z\" (UniqueName: \"kubernetes.io/projected/4aac7e45-016c-4fa2-b383-cb06b4b843d2-kube-api-access-lmt6z\") on node \"crc\" DevicePath \"\"" Dec 11 09:11:20 crc kubenswrapper[4629]: I1211 09:11:20.946952 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aac7e45-016c-4fa2-b383-cb06b4b843d2-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:11:20 crc kubenswrapper[4629]: I1211 09:11:20.965618 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aac7e45-016c-4fa2-b383-cb06b4b843d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4aac7e45-016c-4fa2-b383-cb06b4b843d2" (UID: "4aac7e45-016c-4fa2-b383-cb06b4b843d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.049281 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aac7e45-016c-4fa2-b383-cb06b4b843d2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.292113 4629 generic.go:334] "Generic (PLEG): container finished" podID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" containerID="684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44" exitCode=0 Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.292158 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fhd7" event={"ID":"4aac7e45-016c-4fa2-b383-cb06b4b843d2","Type":"ContainerDied","Data":"684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44"} Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.292191 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fhd7" event={"ID":"4aac7e45-016c-4fa2-b383-cb06b4b843d2","Type":"ContainerDied","Data":"0b5f9025adbe04a306941e8bce9b615a9a87ef950f16905bec9d4a73ec429d6c"} Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.292208 4629 scope.go:117] "RemoveContainer" containerID="684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44" Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.293128 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7fhd7" Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.657907 4629 scope.go:117] "RemoveContainer" containerID="374e4709cc1f3df211673e30472bc0de3d0a4178749a6fc9d00eb45a255598b3" Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.701428 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7fhd7"] Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.723557 4629 scope.go:117] "RemoveContainer" containerID="b6b6e6987387743d1feb4caad2ef01141c57a09bbd81e56e4fb92613a7e69dad" Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.743476 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7fhd7"] Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.799069 4629 scope.go:117] "RemoveContainer" containerID="684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44" Dec 11 09:11:21 crc kubenswrapper[4629]: E1211 09:11:21.800002 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44\": container with ID starting with 684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44 not found: ID does not exist" containerID="684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44" Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.800038 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44"} err="failed to get container status \"684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44\": rpc error: code = NotFound desc = could not find container \"684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44\": container with ID starting with 684784407df7480741ea82693dc4b43359743f4a5028b9afc523bfc2ba07bf44 not found: ID does not exist" Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.800067 4629 scope.go:117] "RemoveContainer" containerID="374e4709cc1f3df211673e30472bc0de3d0a4178749a6fc9d00eb45a255598b3" Dec 11 09:11:21 crc kubenswrapper[4629]: E1211 09:11:21.800335 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"374e4709cc1f3df211673e30472bc0de3d0a4178749a6fc9d00eb45a255598b3\": container with ID starting with 374e4709cc1f3df211673e30472bc0de3d0a4178749a6fc9d00eb45a255598b3 not found: ID does not exist" containerID="374e4709cc1f3df211673e30472bc0de3d0a4178749a6fc9d00eb45a255598b3" Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.800356 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"374e4709cc1f3df211673e30472bc0de3d0a4178749a6fc9d00eb45a255598b3"} err="failed to get container status \"374e4709cc1f3df211673e30472bc0de3d0a4178749a6fc9d00eb45a255598b3\": rpc error: code = NotFound desc = could not find container \"374e4709cc1f3df211673e30472bc0de3d0a4178749a6fc9d00eb45a255598b3\": container with ID starting with 374e4709cc1f3df211673e30472bc0de3d0a4178749a6fc9d00eb45a255598b3 not found: ID does not exist" Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.800373 4629 scope.go:117] "RemoveContainer" containerID="b6b6e6987387743d1feb4caad2ef01141c57a09bbd81e56e4fb92613a7e69dad" Dec 11 09:11:21 crc kubenswrapper[4629]: E1211 09:11:21.800607 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6b6e6987387743d1feb4caad2ef01141c57a09bbd81e56e4fb92613a7e69dad\": container with ID starting with b6b6e6987387743d1feb4caad2ef01141c57a09bbd81e56e4fb92613a7e69dad not found: ID does not exist" containerID="b6b6e6987387743d1feb4caad2ef01141c57a09bbd81e56e4fb92613a7e69dad" Dec 11 09:11:21 crc kubenswrapper[4629]: I1211 09:11:21.800628 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6b6e6987387743d1feb4caad2ef01141c57a09bbd81e56e4fb92613a7e69dad"} err="failed to get container status \"b6b6e6987387743d1feb4caad2ef01141c57a09bbd81e56e4fb92613a7e69dad\": rpc error: code = NotFound desc = could not find container \"b6b6e6987387743d1feb4caad2ef01141c57a09bbd81e56e4fb92613a7e69dad\": container with ID starting with b6b6e6987387743d1feb4caad2ef01141c57a09bbd81e56e4fb92613a7e69dad not found: ID does not exist" Dec 11 09:11:22 crc kubenswrapper[4629]: I1211 09:11:22.213460 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" path="/var/lib/kubelet/pods/4aac7e45-016c-4fa2-b383-cb06b4b843d2/volumes" Dec 11 09:11:50 crc kubenswrapper[4629]: I1211 09:11:50.485972 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:11:50 crc kubenswrapper[4629]: I1211 09:11:50.486578 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:12:20 crc kubenswrapper[4629]: I1211 09:12:20.486057 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:12:20 crc kubenswrapper[4629]: I1211 09:12:20.486788 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:12:20 crc kubenswrapper[4629]: I1211 09:12:20.486890 4629 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 09:12:20 crc kubenswrapper[4629]: I1211 09:12:20.487786 4629 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3070d82dc8c3e310ce3a87a010814089bcbb734e3b66f41c4a60ff1e08bb1f40"} pod="openshift-machine-config-operator/machine-config-daemon-wbld9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 09:12:20 crc kubenswrapper[4629]: I1211 09:12:20.487854 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" containerID="cri-o://3070d82dc8c3e310ce3a87a010814089bcbb734e3b66f41c4a60ff1e08bb1f40" gracePeriod=600 Dec 11 09:12:20 crc kubenswrapper[4629]: I1211 09:12:20.796116 4629 generic.go:334] "Generic (PLEG): container finished" podID="847b50b9-0093-404f-b0dd-63c814d8986b" containerID="3070d82dc8c3e310ce3a87a010814089bcbb734e3b66f41c4a60ff1e08bb1f40" exitCode=0 Dec 11 09:12:20 crc kubenswrapper[4629]: I1211 09:12:20.796161 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerDied","Data":"3070d82dc8c3e310ce3a87a010814089bcbb734e3b66f41c4a60ff1e08bb1f40"} Dec 11 09:12:20 crc kubenswrapper[4629]: I1211 09:12:20.796202 4629 scope.go:117] "RemoveContainer" containerID="099f7736dfb0701582e4848b64e886243f202bb8d67fd4392b4c71e1b1728e3f" Dec 11 09:12:21 crc kubenswrapper[4629]: I1211 09:12:21.806041 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48"} Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.120433 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gs2wr"] Dec 11 09:13:06 crc kubenswrapper[4629]: E1211 09:13:06.122907 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" containerName="extract-utilities" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.122937 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" containerName="extract-utilities" Dec 11 09:13:06 crc kubenswrapper[4629]: E1211 09:13:06.122967 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" containerName="registry-server" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.122976 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" containerName="registry-server" Dec 11 09:13:06 crc kubenswrapper[4629]: E1211 09:13:06.123009 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.123020 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 11 09:13:06 crc kubenswrapper[4629]: E1211 09:13:06.123041 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" containerName="extract-content" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.123049 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" containerName="extract-content" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.123403 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aac7e45-016c-4fa2-b383-cb06b4b843d2" containerName="registry-server" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.123450 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.129257 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.143824 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gs2wr"] Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.215072 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730c1c18-96b4-4d22-a636-2dc8de16a460-utilities\") pod \"certified-operators-gs2wr\" (UID: \"730c1c18-96b4-4d22-a636-2dc8de16a460\") " pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.215216 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730c1c18-96b4-4d22-a636-2dc8de16a460-catalog-content\") pod \"certified-operators-gs2wr\" (UID: \"730c1c18-96b4-4d22-a636-2dc8de16a460\") " pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.215272 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz9ww\" (UniqueName: \"kubernetes.io/projected/730c1c18-96b4-4d22-a636-2dc8de16a460-kube-api-access-cz9ww\") pod \"certified-operators-gs2wr\" (UID: \"730c1c18-96b4-4d22-a636-2dc8de16a460\") " pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.316920 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730c1c18-96b4-4d22-a636-2dc8de16a460-catalog-content\") pod \"certified-operators-gs2wr\" (UID: \"730c1c18-96b4-4d22-a636-2dc8de16a460\") " pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.317016 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz9ww\" (UniqueName: \"kubernetes.io/projected/730c1c18-96b4-4d22-a636-2dc8de16a460-kube-api-access-cz9ww\") pod \"certified-operators-gs2wr\" (UID: \"730c1c18-96b4-4d22-a636-2dc8de16a460\") " pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.317141 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730c1c18-96b4-4d22-a636-2dc8de16a460-utilities\") pod \"certified-operators-gs2wr\" (UID: \"730c1c18-96b4-4d22-a636-2dc8de16a460\") " pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.317880 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730c1c18-96b4-4d22-a636-2dc8de16a460-catalog-content\") pod \"certified-operators-gs2wr\" (UID: \"730c1c18-96b4-4d22-a636-2dc8de16a460\") " pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.317951 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730c1c18-96b4-4d22-a636-2dc8de16a460-utilities\") pod \"certified-operators-gs2wr\" (UID: \"730c1c18-96b4-4d22-a636-2dc8de16a460\") " pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.339796 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz9ww\" (UniqueName: \"kubernetes.io/projected/730c1c18-96b4-4d22-a636-2dc8de16a460-kube-api-access-cz9ww\") pod \"certified-operators-gs2wr\" (UID: \"730c1c18-96b4-4d22-a636-2dc8de16a460\") " pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:06 crc kubenswrapper[4629]: I1211 09:13:06.461686 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:07 crc kubenswrapper[4629]: I1211 09:13:07.081819 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gs2wr"] Dec 11 09:13:07 crc kubenswrapper[4629]: I1211 09:13:07.322130 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2wr" event={"ID":"730c1c18-96b4-4d22-a636-2dc8de16a460","Type":"ContainerStarted","Data":"1107343228e0e9d05e2587736bf30b6ad425f6e096824b4f7d7291935628360b"} Dec 11 09:13:08 crc kubenswrapper[4629]: I1211 09:13:08.334133 4629 generic.go:334] "Generic (PLEG): container finished" podID="730c1c18-96b4-4d22-a636-2dc8de16a460" containerID="ee3fa2d810a902772caeda5f9bc0dbcd0cc9b924d6da0ffff0a0bedd32746201" exitCode=0 Dec 11 09:13:08 crc kubenswrapper[4629]: I1211 09:13:08.334357 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2wr" event={"ID":"730c1c18-96b4-4d22-a636-2dc8de16a460","Type":"ContainerDied","Data":"ee3fa2d810a902772caeda5f9bc0dbcd0cc9b924d6da0ffff0a0bedd32746201"} Dec 11 09:13:08 crc kubenswrapper[4629]: I1211 09:13:08.336980 4629 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 09:13:09 crc kubenswrapper[4629]: I1211 09:13:09.344362 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2wr" event={"ID":"730c1c18-96b4-4d22-a636-2dc8de16a460","Type":"ContainerStarted","Data":"6ff21bddee10ad4903a4cf9363fee76dcdbcb94ca552efce4ebc6caa7c8297ee"} Dec 11 09:13:10 crc kubenswrapper[4629]: I1211 09:13:10.354989 4629 generic.go:334] "Generic (PLEG): container finished" podID="730c1c18-96b4-4d22-a636-2dc8de16a460" containerID="6ff21bddee10ad4903a4cf9363fee76dcdbcb94ca552efce4ebc6caa7c8297ee" exitCode=0 Dec 11 09:13:10 crc kubenswrapper[4629]: I1211 09:13:10.355043 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2wr" event={"ID":"730c1c18-96b4-4d22-a636-2dc8de16a460","Type":"ContainerDied","Data":"6ff21bddee10ad4903a4cf9363fee76dcdbcb94ca552efce4ebc6caa7c8297ee"} Dec 11 09:13:12 crc kubenswrapper[4629]: I1211 09:13:12.378361 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2wr" event={"ID":"730c1c18-96b4-4d22-a636-2dc8de16a460","Type":"ContainerStarted","Data":"43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1"} Dec 11 09:13:12 crc kubenswrapper[4629]: I1211 09:13:12.407653 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gs2wr" podStartSLOduration=2.897951812 podStartE2EDuration="6.407613215s" podCreationTimestamp="2025-12-11 09:13:06 +0000 UTC" firstStartedPulling="2025-12-11 09:13:08.336647755 +0000 UTC m=+2176.402063363" lastFinishedPulling="2025-12-11 09:13:11.846309158 +0000 UTC m=+2179.911724766" observedRunningTime="2025-12-11 09:13:12.395612857 +0000 UTC m=+2180.461028475" watchObservedRunningTime="2025-12-11 09:13:12.407613215 +0000 UTC m=+2180.473028813" Dec 11 09:13:16 crc kubenswrapper[4629]: I1211 09:13:16.462158 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:16 crc kubenswrapper[4629]: I1211 09:13:16.462795 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:16 crc kubenswrapper[4629]: I1211 09:13:16.508410 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:17 crc kubenswrapper[4629]: I1211 09:13:17.510578 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:17 crc kubenswrapper[4629]: I1211 09:13:17.567574 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gs2wr"] Dec 11 09:13:19 crc kubenswrapper[4629]: I1211 09:13:19.467057 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gs2wr" podUID="730c1c18-96b4-4d22-a636-2dc8de16a460" containerName="registry-server" containerID="cri-o://43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1" gracePeriod=2 Dec 11 09:13:19 crc kubenswrapper[4629]: I1211 09:13:19.925043 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.027944 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730c1c18-96b4-4d22-a636-2dc8de16a460-catalog-content\") pod \"730c1c18-96b4-4d22-a636-2dc8de16a460\" (UID: \"730c1c18-96b4-4d22-a636-2dc8de16a460\") " Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.028066 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz9ww\" (UniqueName: \"kubernetes.io/projected/730c1c18-96b4-4d22-a636-2dc8de16a460-kube-api-access-cz9ww\") pod \"730c1c18-96b4-4d22-a636-2dc8de16a460\" (UID: \"730c1c18-96b4-4d22-a636-2dc8de16a460\") " Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.028223 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730c1c18-96b4-4d22-a636-2dc8de16a460-utilities\") pod \"730c1c18-96b4-4d22-a636-2dc8de16a460\" (UID: \"730c1c18-96b4-4d22-a636-2dc8de16a460\") " Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.029739 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/730c1c18-96b4-4d22-a636-2dc8de16a460-utilities" (OuterVolumeSpecName: "utilities") pod "730c1c18-96b4-4d22-a636-2dc8de16a460" (UID: "730c1c18-96b4-4d22-a636-2dc8de16a460"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.039234 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/730c1c18-96b4-4d22-a636-2dc8de16a460-kube-api-access-cz9ww" (OuterVolumeSpecName: "kube-api-access-cz9ww") pod "730c1c18-96b4-4d22-a636-2dc8de16a460" (UID: "730c1c18-96b4-4d22-a636-2dc8de16a460"). InnerVolumeSpecName "kube-api-access-cz9ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.085710 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/730c1c18-96b4-4d22-a636-2dc8de16a460-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "730c1c18-96b4-4d22-a636-2dc8de16a460" (UID: "730c1c18-96b4-4d22-a636-2dc8de16a460"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.130673 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730c1c18-96b4-4d22-a636-2dc8de16a460-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.130723 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz9ww\" (UniqueName: \"kubernetes.io/projected/730c1c18-96b4-4d22-a636-2dc8de16a460-kube-api-access-cz9ww\") on node \"crc\" DevicePath \"\"" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.130739 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730c1c18-96b4-4d22-a636-2dc8de16a460-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.479920 4629 generic.go:334] "Generic (PLEG): container finished" podID="730c1c18-96b4-4d22-a636-2dc8de16a460" containerID="43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1" exitCode=0 Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.479967 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2wr" event={"ID":"730c1c18-96b4-4d22-a636-2dc8de16a460","Type":"ContainerDied","Data":"43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1"} Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.479995 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2wr" event={"ID":"730c1c18-96b4-4d22-a636-2dc8de16a460","Type":"ContainerDied","Data":"1107343228e0e9d05e2587736bf30b6ad425f6e096824b4f7d7291935628360b"} Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.480011 4629 scope.go:117] "RemoveContainer" containerID="43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.480140 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gs2wr" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.508255 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gs2wr"] Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.508709 4629 scope.go:117] "RemoveContainer" containerID="6ff21bddee10ad4903a4cf9363fee76dcdbcb94ca552efce4ebc6caa7c8297ee" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.516481 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gs2wr"] Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.545471 4629 scope.go:117] "RemoveContainer" containerID="ee3fa2d810a902772caeda5f9bc0dbcd0cc9b924d6da0ffff0a0bedd32746201" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.570826 4629 scope.go:117] "RemoveContainer" containerID="43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1" Dec 11 09:13:20 crc kubenswrapper[4629]: E1211 09:13:20.571396 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1\": container with ID starting with 43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1 not found: ID does not exist" containerID="43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.571439 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1"} err="failed to get container status \"43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1\": rpc error: code = NotFound desc = could not find container \"43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1\": container with ID starting with 43771a17706450e6220a221a63c586178b4cb14a44bd2b0ccfef1bd65bc93cd1 not found: ID does not exist" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.571479 4629 scope.go:117] "RemoveContainer" containerID="6ff21bddee10ad4903a4cf9363fee76dcdbcb94ca552efce4ebc6caa7c8297ee" Dec 11 09:13:20 crc kubenswrapper[4629]: E1211 09:13:20.571822 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ff21bddee10ad4903a4cf9363fee76dcdbcb94ca552efce4ebc6caa7c8297ee\": container with ID starting with 6ff21bddee10ad4903a4cf9363fee76dcdbcb94ca552efce4ebc6caa7c8297ee not found: ID does not exist" containerID="6ff21bddee10ad4903a4cf9363fee76dcdbcb94ca552efce4ebc6caa7c8297ee" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.571877 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ff21bddee10ad4903a4cf9363fee76dcdbcb94ca552efce4ebc6caa7c8297ee"} err="failed to get container status \"6ff21bddee10ad4903a4cf9363fee76dcdbcb94ca552efce4ebc6caa7c8297ee\": rpc error: code = NotFound desc = could not find container \"6ff21bddee10ad4903a4cf9363fee76dcdbcb94ca552efce4ebc6caa7c8297ee\": container with ID starting with 6ff21bddee10ad4903a4cf9363fee76dcdbcb94ca552efce4ebc6caa7c8297ee not found: ID does not exist" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.571911 4629 scope.go:117] "RemoveContainer" containerID="ee3fa2d810a902772caeda5f9bc0dbcd0cc9b924d6da0ffff0a0bedd32746201" Dec 11 09:13:20 crc kubenswrapper[4629]: E1211 09:13:20.572207 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee3fa2d810a902772caeda5f9bc0dbcd0cc9b924d6da0ffff0a0bedd32746201\": container with ID starting with ee3fa2d810a902772caeda5f9bc0dbcd0cc9b924d6da0ffff0a0bedd32746201 not found: ID does not exist" containerID="ee3fa2d810a902772caeda5f9bc0dbcd0cc9b924d6da0ffff0a0bedd32746201" Dec 11 09:13:20 crc kubenswrapper[4629]: I1211 09:13:20.572232 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee3fa2d810a902772caeda5f9bc0dbcd0cc9b924d6da0ffff0a0bedd32746201"} err="failed to get container status \"ee3fa2d810a902772caeda5f9bc0dbcd0cc9b924d6da0ffff0a0bedd32746201\": rpc error: code = NotFound desc = could not find container \"ee3fa2d810a902772caeda5f9bc0dbcd0cc9b924d6da0ffff0a0bedd32746201\": container with ID starting with ee3fa2d810a902772caeda5f9bc0dbcd0cc9b924d6da0ffff0a0bedd32746201 not found: ID does not exist" Dec 11 09:13:22 crc kubenswrapper[4629]: I1211 09:13:22.232275 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="730c1c18-96b4-4d22-a636-2dc8de16a460" path="/var/lib/kubelet/pods/730c1c18-96b4-4d22-a636-2dc8de16a460/volumes" Dec 11 09:13:28 crc kubenswrapper[4629]: I1211 09:13:28.838347 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rc25k"] Dec 11 09:13:28 crc kubenswrapper[4629]: E1211 09:13:28.839253 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="730c1c18-96b4-4d22-a636-2dc8de16a460" containerName="extract-content" Dec 11 09:13:28 crc kubenswrapper[4629]: I1211 09:13:28.839266 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="730c1c18-96b4-4d22-a636-2dc8de16a460" containerName="extract-content" Dec 11 09:13:28 crc kubenswrapper[4629]: E1211 09:13:28.839277 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="730c1c18-96b4-4d22-a636-2dc8de16a460" containerName="extract-utilities" Dec 11 09:13:28 crc kubenswrapper[4629]: I1211 09:13:28.839283 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="730c1c18-96b4-4d22-a636-2dc8de16a460" containerName="extract-utilities" Dec 11 09:13:28 crc kubenswrapper[4629]: E1211 09:13:28.839301 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="730c1c18-96b4-4d22-a636-2dc8de16a460" containerName="registry-server" Dec 11 09:13:28 crc kubenswrapper[4629]: I1211 09:13:28.839307 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="730c1c18-96b4-4d22-a636-2dc8de16a460" containerName="registry-server" Dec 11 09:13:28 crc kubenswrapper[4629]: I1211 09:13:28.839469 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="730c1c18-96b4-4d22-a636-2dc8de16a460" containerName="registry-server" Dec 11 09:13:28 crc kubenswrapper[4629]: I1211 09:13:28.840831 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:28 crc kubenswrapper[4629]: I1211 09:13:28.850363 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rc25k"] Dec 11 09:13:28 crc kubenswrapper[4629]: I1211 09:13:28.935625 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-catalog-content\") pod \"redhat-marketplace-rc25k\" (UID: \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\") " pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:28 crc kubenswrapper[4629]: I1211 09:13:28.936172 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-utilities\") pod \"redhat-marketplace-rc25k\" (UID: \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\") " pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:28 crc kubenswrapper[4629]: I1211 09:13:28.936348 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9s4g\" (UniqueName: \"kubernetes.io/projected/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-kube-api-access-g9s4g\") pod \"redhat-marketplace-rc25k\" (UID: \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\") " pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:29 crc kubenswrapper[4629]: I1211 09:13:29.038597 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-catalog-content\") pod \"redhat-marketplace-rc25k\" (UID: \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\") " pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:29 crc kubenswrapper[4629]: I1211 09:13:29.038948 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-utilities\") pod \"redhat-marketplace-rc25k\" (UID: \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\") " pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:29 crc kubenswrapper[4629]: I1211 09:13:29.039050 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9s4g\" (UniqueName: \"kubernetes.io/projected/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-kube-api-access-g9s4g\") pod \"redhat-marketplace-rc25k\" (UID: \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\") " pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:29 crc kubenswrapper[4629]: I1211 09:13:29.039641 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-catalog-content\") pod \"redhat-marketplace-rc25k\" (UID: \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\") " pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:29 crc kubenswrapper[4629]: I1211 09:13:29.039707 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-utilities\") pod \"redhat-marketplace-rc25k\" (UID: \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\") " pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:29 crc kubenswrapper[4629]: I1211 09:13:29.063446 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9s4g\" (UniqueName: \"kubernetes.io/projected/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-kube-api-access-g9s4g\") pod \"redhat-marketplace-rc25k\" (UID: \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\") " pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:29 crc kubenswrapper[4629]: I1211 09:13:29.158690 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:29 crc kubenswrapper[4629]: I1211 09:13:29.683727 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rc25k"] Dec 11 09:13:29 crc kubenswrapper[4629]: W1211 09:13:29.688945 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c8460ef_7c95_4963_ab7f_b4fe0eec2eea.slice/crio-75bf1f11b8c99b9d1c05b39943e6b5644947dd3aa90a878527fc6b287d9ba770 WatchSource:0}: Error finding container 75bf1f11b8c99b9d1c05b39943e6b5644947dd3aa90a878527fc6b287d9ba770: Status 404 returned error can't find the container with id 75bf1f11b8c99b9d1c05b39943e6b5644947dd3aa90a878527fc6b287d9ba770 Dec 11 09:13:30 crc kubenswrapper[4629]: I1211 09:13:30.556131 4629 generic.go:334] "Generic (PLEG): container finished" podID="4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" containerID="b37e1c68025db375e11b6ec9e7fbd374a57651449525ce83d6cf456e76dfc055" exitCode=0 Dec 11 09:13:30 crc kubenswrapper[4629]: I1211 09:13:30.556194 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rc25k" event={"ID":"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea","Type":"ContainerDied","Data":"b37e1c68025db375e11b6ec9e7fbd374a57651449525ce83d6cf456e76dfc055"} Dec 11 09:13:30 crc kubenswrapper[4629]: I1211 09:13:30.556251 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rc25k" event={"ID":"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea","Type":"ContainerStarted","Data":"75bf1f11b8c99b9d1c05b39943e6b5644947dd3aa90a878527fc6b287d9ba770"} Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.237227 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8btjh"] Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.240501 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.258707 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8btjh"] Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.430837 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-utilities\") pod \"community-operators-8btjh\" (UID: \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\") " pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.430922 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-catalog-content\") pod \"community-operators-8btjh\" (UID: \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\") " pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.431086 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2fm6\" (UniqueName: \"kubernetes.io/projected/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-kube-api-access-k2fm6\") pod \"community-operators-8btjh\" (UID: \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\") " pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.533660 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-utilities\") pod \"community-operators-8btjh\" (UID: \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\") " pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.534245 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-utilities\") pod \"community-operators-8btjh\" (UID: \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\") " pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.535204 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-catalog-content\") pod \"community-operators-8btjh\" (UID: \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\") " pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.535472 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-catalog-content\") pod \"community-operators-8btjh\" (UID: \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\") " pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.535710 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2fm6\" (UniqueName: \"kubernetes.io/projected/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-kube-api-access-k2fm6\") pod \"community-operators-8btjh\" (UID: \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\") " pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.556659 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2fm6\" (UniqueName: \"kubernetes.io/projected/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-kube-api-access-k2fm6\") pod \"community-operators-8btjh\" (UID: \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\") " pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.560337 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.585993 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rc25k" event={"ID":"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea","Type":"ContainerStarted","Data":"90a42d2c9dff8b0cef601b18c27a4ca587f723a18eb867c0b69cdff8dd3388bb"} Dec 11 09:13:33 crc kubenswrapper[4629]: I1211 09:13:33.957631 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8btjh"] Dec 11 09:13:34 crc kubenswrapper[4629]: I1211 09:13:34.596962 4629 generic.go:334] "Generic (PLEG): container finished" podID="4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" containerID="90a42d2c9dff8b0cef601b18c27a4ca587f723a18eb867c0b69cdff8dd3388bb" exitCode=0 Dec 11 09:13:34 crc kubenswrapper[4629]: I1211 09:13:34.597033 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rc25k" event={"ID":"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea","Type":"ContainerDied","Data":"90a42d2c9dff8b0cef601b18c27a4ca587f723a18eb867c0b69cdff8dd3388bb"} Dec 11 09:13:34 crc kubenswrapper[4629]: I1211 09:13:34.600354 4629 generic.go:334] "Generic (PLEG): container finished" podID="bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" containerID="4eea57cb55f2a0f6e35eefde2a1a390211d5b0409bb57f5fd7822d5ca9b8467f" exitCode=0 Dec 11 09:13:34 crc kubenswrapper[4629]: I1211 09:13:34.600405 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8btjh" event={"ID":"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b","Type":"ContainerDied","Data":"4eea57cb55f2a0f6e35eefde2a1a390211d5b0409bb57f5fd7822d5ca9b8467f"} Dec 11 09:13:34 crc kubenswrapper[4629]: I1211 09:13:34.600437 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8btjh" event={"ID":"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b","Type":"ContainerStarted","Data":"19201b1d314faaeefe247064ebdab849e964e08533e2be9db6fe6e5ea5643e40"} Dec 11 09:13:35 crc kubenswrapper[4629]: I1211 09:13:35.614474 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rc25k" event={"ID":"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea","Type":"ContainerStarted","Data":"7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392"} Dec 11 09:13:35 crc kubenswrapper[4629]: I1211 09:13:35.633705 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rc25k" podStartSLOduration=3.043667294 podStartE2EDuration="7.633683031s" podCreationTimestamp="2025-12-11 09:13:28 +0000 UTC" firstStartedPulling="2025-12-11 09:13:30.558720331 +0000 UTC m=+2198.624135949" lastFinishedPulling="2025-12-11 09:13:35.148736078 +0000 UTC m=+2203.214151686" observedRunningTime="2025-12-11 09:13:35.632544496 +0000 UTC m=+2203.697960114" watchObservedRunningTime="2025-12-11 09:13:35.633683031 +0000 UTC m=+2203.699098639" Dec 11 09:13:36 crc kubenswrapper[4629]: I1211 09:13:36.624258 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8btjh" event={"ID":"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b","Type":"ContainerStarted","Data":"605dfbf49a9c53d20ec1aa4f58554243b781fa7c47a5d8f83762facc1993acb0"} Dec 11 09:13:37 crc kubenswrapper[4629]: I1211 09:13:37.632965 4629 generic.go:334] "Generic (PLEG): container finished" podID="bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" containerID="605dfbf49a9c53d20ec1aa4f58554243b781fa7c47a5d8f83762facc1993acb0" exitCode=0 Dec 11 09:13:37 crc kubenswrapper[4629]: I1211 09:13:37.633196 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8btjh" event={"ID":"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b","Type":"ContainerDied","Data":"605dfbf49a9c53d20ec1aa4f58554243b781fa7c47a5d8f83762facc1993acb0"} Dec 11 09:13:38 crc kubenswrapper[4629]: I1211 09:13:38.644481 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8btjh" event={"ID":"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b","Type":"ContainerStarted","Data":"0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05"} Dec 11 09:13:38 crc kubenswrapper[4629]: I1211 09:13:38.664089 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8btjh" podStartSLOduration=2.013474128 podStartE2EDuration="5.664070757s" podCreationTimestamp="2025-12-11 09:13:33 +0000 UTC" firstStartedPulling="2025-12-11 09:13:34.60223861 +0000 UTC m=+2202.667654218" lastFinishedPulling="2025-12-11 09:13:38.252835239 +0000 UTC m=+2206.318250847" observedRunningTime="2025-12-11 09:13:38.661787805 +0000 UTC m=+2206.727203413" watchObservedRunningTime="2025-12-11 09:13:38.664070757 +0000 UTC m=+2206.729486365" Dec 11 09:13:39 crc kubenswrapper[4629]: I1211 09:13:39.160031 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:39 crc kubenswrapper[4629]: I1211 09:13:39.160104 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:39 crc kubenswrapper[4629]: I1211 09:13:39.211358 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:43 crc kubenswrapper[4629]: I1211 09:13:43.561533 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:43 crc kubenswrapper[4629]: I1211 09:13:43.562360 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:43 crc kubenswrapper[4629]: I1211 09:13:43.611328 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:43 crc kubenswrapper[4629]: I1211 09:13:43.740288 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:44 crc kubenswrapper[4629]: I1211 09:13:44.422137 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8btjh"] Dec 11 09:13:45 crc kubenswrapper[4629]: I1211 09:13:45.713937 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8btjh" podUID="bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" containerName="registry-server" containerID="cri-o://0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05" gracePeriod=2 Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.142970 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.276512 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2fm6\" (UniqueName: \"kubernetes.io/projected/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-kube-api-access-k2fm6\") pod \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\" (UID: \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\") " Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.276647 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-utilities\") pod \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\" (UID: \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\") " Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.276802 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-catalog-content\") pod \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\" (UID: \"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b\") " Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.279405 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-utilities" (OuterVolumeSpecName: "utilities") pod "bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" (UID: "bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.284131 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-kube-api-access-k2fm6" (OuterVolumeSpecName: "kube-api-access-k2fm6") pod "bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" (UID: "bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b"). InnerVolumeSpecName "kube-api-access-k2fm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.328134 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" (UID: "bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.379122 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2fm6\" (UniqueName: \"kubernetes.io/projected/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-kube-api-access-k2fm6\") on node \"crc\" DevicePath \"\"" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.379350 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.379441 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.722703 4629 generic.go:334] "Generic (PLEG): container finished" podID="bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" containerID="0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05" exitCode=0 Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.722742 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8btjh" event={"ID":"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b","Type":"ContainerDied","Data":"0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05"} Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.722770 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8btjh" event={"ID":"bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b","Type":"ContainerDied","Data":"19201b1d314faaeefe247064ebdab849e964e08533e2be9db6fe6e5ea5643e40"} Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.722785 4629 scope.go:117] "RemoveContainer" containerID="0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.722930 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8btjh" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.749858 4629 scope.go:117] "RemoveContainer" containerID="605dfbf49a9c53d20ec1aa4f58554243b781fa7c47a5d8f83762facc1993acb0" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.760400 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8btjh"] Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.767755 4629 scope.go:117] "RemoveContainer" containerID="4eea57cb55f2a0f6e35eefde2a1a390211d5b0409bb57f5fd7822d5ca9b8467f" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.770803 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8btjh"] Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.814699 4629 scope.go:117] "RemoveContainer" containerID="0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05" Dec 11 09:13:46 crc kubenswrapper[4629]: E1211 09:13:46.815713 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05\": container with ID starting with 0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05 not found: ID does not exist" containerID="0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.815775 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05"} err="failed to get container status \"0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05\": rpc error: code = NotFound desc = could not find container \"0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05\": container with ID starting with 0eb20785855140fe7b2967d6e04a79e7cee260dca5a216520227c8cef14aaa05 not found: ID does not exist" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.815809 4629 scope.go:117] "RemoveContainer" containerID="605dfbf49a9c53d20ec1aa4f58554243b781fa7c47a5d8f83762facc1993acb0" Dec 11 09:13:46 crc kubenswrapper[4629]: E1211 09:13:46.817253 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"605dfbf49a9c53d20ec1aa4f58554243b781fa7c47a5d8f83762facc1993acb0\": container with ID starting with 605dfbf49a9c53d20ec1aa4f58554243b781fa7c47a5d8f83762facc1993acb0 not found: ID does not exist" containerID="605dfbf49a9c53d20ec1aa4f58554243b781fa7c47a5d8f83762facc1993acb0" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.817289 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"605dfbf49a9c53d20ec1aa4f58554243b781fa7c47a5d8f83762facc1993acb0"} err="failed to get container status \"605dfbf49a9c53d20ec1aa4f58554243b781fa7c47a5d8f83762facc1993acb0\": rpc error: code = NotFound desc = could not find container \"605dfbf49a9c53d20ec1aa4f58554243b781fa7c47a5d8f83762facc1993acb0\": container with ID starting with 605dfbf49a9c53d20ec1aa4f58554243b781fa7c47a5d8f83762facc1993acb0 not found: ID does not exist" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.817305 4629 scope.go:117] "RemoveContainer" containerID="4eea57cb55f2a0f6e35eefde2a1a390211d5b0409bb57f5fd7822d5ca9b8467f" Dec 11 09:13:46 crc kubenswrapper[4629]: E1211 09:13:46.817690 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eea57cb55f2a0f6e35eefde2a1a390211d5b0409bb57f5fd7822d5ca9b8467f\": container with ID starting with 4eea57cb55f2a0f6e35eefde2a1a390211d5b0409bb57f5fd7822d5ca9b8467f not found: ID does not exist" containerID="4eea57cb55f2a0f6e35eefde2a1a390211d5b0409bb57f5fd7822d5ca9b8467f" Dec 11 09:13:46 crc kubenswrapper[4629]: I1211 09:13:46.817713 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eea57cb55f2a0f6e35eefde2a1a390211d5b0409bb57f5fd7822d5ca9b8467f"} err="failed to get container status \"4eea57cb55f2a0f6e35eefde2a1a390211d5b0409bb57f5fd7822d5ca9b8467f\": rpc error: code = NotFound desc = could not find container \"4eea57cb55f2a0f6e35eefde2a1a390211d5b0409bb57f5fd7822d5ca9b8467f\": container with ID starting with 4eea57cb55f2a0f6e35eefde2a1a390211d5b0409bb57f5fd7822d5ca9b8467f not found: ID does not exist" Dec 11 09:13:48 crc kubenswrapper[4629]: I1211 09:13:48.210316 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" path="/var/lib/kubelet/pods/bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b/volumes" Dec 11 09:13:49 crc kubenswrapper[4629]: I1211 09:13:49.209454 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:49 crc kubenswrapper[4629]: I1211 09:13:49.824135 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rc25k"] Dec 11 09:13:49 crc kubenswrapper[4629]: I1211 09:13:49.824438 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rc25k" podUID="4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" containerName="registry-server" containerID="cri-o://7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392" gracePeriod=2 Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.301395 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.357327 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-catalog-content\") pod \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\" (UID: \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\") " Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.357380 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-utilities\") pod \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\" (UID: \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\") " Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.357422 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9s4g\" (UniqueName: \"kubernetes.io/projected/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-kube-api-access-g9s4g\") pod \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\" (UID: \"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea\") " Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.358358 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-utilities" (OuterVolumeSpecName: "utilities") pod "4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" (UID: "4c8460ef-7c95-4963-ab7f-b4fe0eec2eea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.358736 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.364352 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-kube-api-access-g9s4g" (OuterVolumeSpecName: "kube-api-access-g9s4g") pod "4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" (UID: "4c8460ef-7c95-4963-ab7f-b4fe0eec2eea"). InnerVolumeSpecName "kube-api-access-g9s4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.377079 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" (UID: "4c8460ef-7c95-4963-ab7f-b4fe0eec2eea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.460477 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9s4g\" (UniqueName: \"kubernetes.io/projected/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-kube-api-access-g9s4g\") on node \"crc\" DevicePath \"\"" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.460513 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.765001 4629 generic.go:334] "Generic (PLEG): container finished" podID="4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" containerID="7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392" exitCode=0 Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.765052 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rc25k" event={"ID":"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea","Type":"ContainerDied","Data":"7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392"} Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.765081 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rc25k" event={"ID":"4c8460ef-7c95-4963-ab7f-b4fe0eec2eea","Type":"ContainerDied","Data":"75bf1f11b8c99b9d1c05b39943e6b5644947dd3aa90a878527fc6b287d9ba770"} Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.765100 4629 scope.go:117] "RemoveContainer" containerID="7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.765225 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rc25k" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.784568 4629 scope.go:117] "RemoveContainer" containerID="90a42d2c9dff8b0cef601b18c27a4ca587f723a18eb867c0b69cdff8dd3388bb" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.801991 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rc25k"] Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.810909 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rc25k"] Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.815320 4629 scope.go:117] "RemoveContainer" containerID="b37e1c68025db375e11b6ec9e7fbd374a57651449525ce83d6cf456e76dfc055" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.868628 4629 scope.go:117] "RemoveContainer" containerID="7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392" Dec 11 09:13:50 crc kubenswrapper[4629]: E1211 09:13:50.869439 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392\": container with ID starting with 7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392 not found: ID does not exist" containerID="7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.869563 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392"} err="failed to get container status \"7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392\": rpc error: code = NotFound desc = could not find container \"7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392\": container with ID starting with 7d907ff02acf4b7d6814b1b2b05dc9ff63826e8e4e22794467fa23bff349d392 not found: ID does not exist" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.869646 4629 scope.go:117] "RemoveContainer" containerID="90a42d2c9dff8b0cef601b18c27a4ca587f723a18eb867c0b69cdff8dd3388bb" Dec 11 09:13:50 crc kubenswrapper[4629]: E1211 09:13:50.869987 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90a42d2c9dff8b0cef601b18c27a4ca587f723a18eb867c0b69cdff8dd3388bb\": container with ID starting with 90a42d2c9dff8b0cef601b18c27a4ca587f723a18eb867c0b69cdff8dd3388bb not found: ID does not exist" containerID="90a42d2c9dff8b0cef601b18c27a4ca587f723a18eb867c0b69cdff8dd3388bb" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.870015 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90a42d2c9dff8b0cef601b18c27a4ca587f723a18eb867c0b69cdff8dd3388bb"} err="failed to get container status \"90a42d2c9dff8b0cef601b18c27a4ca587f723a18eb867c0b69cdff8dd3388bb\": rpc error: code = NotFound desc = could not find container \"90a42d2c9dff8b0cef601b18c27a4ca587f723a18eb867c0b69cdff8dd3388bb\": container with ID starting with 90a42d2c9dff8b0cef601b18c27a4ca587f723a18eb867c0b69cdff8dd3388bb not found: ID does not exist" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.870029 4629 scope.go:117] "RemoveContainer" containerID="b37e1c68025db375e11b6ec9e7fbd374a57651449525ce83d6cf456e76dfc055" Dec 11 09:13:50 crc kubenswrapper[4629]: E1211 09:13:50.870327 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b37e1c68025db375e11b6ec9e7fbd374a57651449525ce83d6cf456e76dfc055\": container with ID starting with b37e1c68025db375e11b6ec9e7fbd374a57651449525ce83d6cf456e76dfc055 not found: ID does not exist" containerID="b37e1c68025db375e11b6ec9e7fbd374a57651449525ce83d6cf456e76dfc055" Dec 11 09:13:50 crc kubenswrapper[4629]: I1211 09:13:50.870356 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b37e1c68025db375e11b6ec9e7fbd374a57651449525ce83d6cf456e76dfc055"} err="failed to get container status \"b37e1c68025db375e11b6ec9e7fbd374a57651449525ce83d6cf456e76dfc055\": rpc error: code = NotFound desc = could not find container \"b37e1c68025db375e11b6ec9e7fbd374a57651449525ce83d6cf456e76dfc055\": container with ID starting with b37e1c68025db375e11b6ec9e7fbd374a57651449525ce83d6cf456e76dfc055 not found: ID does not exist" Dec 11 09:13:52 crc kubenswrapper[4629]: I1211 09:13:52.217262 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" path="/var/lib/kubelet/pods/4c8460ef-7c95-4963-ab7f-b4fe0eec2eea/volumes" Dec 11 09:14:50 crc kubenswrapper[4629]: I1211 09:14:50.486597 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:14:50 crc kubenswrapper[4629]: I1211 09:14:50.487260 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.230762 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf"] Dec 11 09:15:00 crc kubenswrapper[4629]: E1211 09:15:00.231981 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" containerName="extract-content" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.232004 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" containerName="extract-content" Dec 11 09:15:00 crc kubenswrapper[4629]: E1211 09:15:00.232041 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" containerName="registry-server" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.232050 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" containerName="registry-server" Dec 11 09:15:00 crc kubenswrapper[4629]: E1211 09:15:00.232065 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" containerName="extract-content" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.232073 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" containerName="extract-content" Dec 11 09:15:00 crc kubenswrapper[4629]: E1211 09:15:00.232092 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" containerName="extract-utilities" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.232099 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" containerName="extract-utilities" Dec 11 09:15:00 crc kubenswrapper[4629]: E1211 09:15:00.232122 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" containerName="extract-utilities" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.232129 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" containerName="extract-utilities" Dec 11 09:15:00 crc kubenswrapper[4629]: E1211 09:15:00.232148 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" containerName="registry-server" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.232155 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" containerName="registry-server" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.232375 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdbb3adf-7aaf-4679-aa2d-d1bc1baefb9b" containerName="registry-server" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.232408 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c8460ef-7c95-4963-ab7f-b4fe0eec2eea" containerName="registry-server" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.233251 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.236371 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.237300 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.240008 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf"] Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.300351 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbkqz\" (UniqueName: \"kubernetes.io/projected/b51b84fa-52d1-41b1-937b-4a5a568fd917-kube-api-access-qbkqz\") pod \"collect-profiles-29424075-mtfjf\" (UID: \"b51b84fa-52d1-41b1-937b-4a5a568fd917\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.300542 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b51b84fa-52d1-41b1-937b-4a5a568fd917-config-volume\") pod \"collect-profiles-29424075-mtfjf\" (UID: \"b51b84fa-52d1-41b1-937b-4a5a568fd917\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.300606 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b51b84fa-52d1-41b1-937b-4a5a568fd917-secret-volume\") pod \"collect-profiles-29424075-mtfjf\" (UID: \"b51b84fa-52d1-41b1-937b-4a5a568fd917\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.402538 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b51b84fa-52d1-41b1-937b-4a5a568fd917-config-volume\") pod \"collect-profiles-29424075-mtfjf\" (UID: \"b51b84fa-52d1-41b1-937b-4a5a568fd917\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.402627 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b51b84fa-52d1-41b1-937b-4a5a568fd917-secret-volume\") pod \"collect-profiles-29424075-mtfjf\" (UID: \"b51b84fa-52d1-41b1-937b-4a5a568fd917\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.402677 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbkqz\" (UniqueName: \"kubernetes.io/projected/b51b84fa-52d1-41b1-937b-4a5a568fd917-kube-api-access-qbkqz\") pod \"collect-profiles-29424075-mtfjf\" (UID: \"b51b84fa-52d1-41b1-937b-4a5a568fd917\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.403767 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b51b84fa-52d1-41b1-937b-4a5a568fd917-config-volume\") pod \"collect-profiles-29424075-mtfjf\" (UID: \"b51b84fa-52d1-41b1-937b-4a5a568fd917\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.410128 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b51b84fa-52d1-41b1-937b-4a5a568fd917-secret-volume\") pod \"collect-profiles-29424075-mtfjf\" (UID: \"b51b84fa-52d1-41b1-937b-4a5a568fd917\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.423888 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbkqz\" (UniqueName: \"kubernetes.io/projected/b51b84fa-52d1-41b1-937b-4a5a568fd917-kube-api-access-qbkqz\") pod \"collect-profiles-29424075-mtfjf\" (UID: \"b51b84fa-52d1-41b1-937b-4a5a568fd917\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:00 crc kubenswrapper[4629]: I1211 09:15:00.560698 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:01 crc kubenswrapper[4629]: I1211 09:15:01.092784 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf"] Dec 11 09:15:01 crc kubenswrapper[4629]: I1211 09:15:01.455878 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" event={"ID":"b51b84fa-52d1-41b1-937b-4a5a568fd917","Type":"ContainerStarted","Data":"03e5fe5e57475704cf77b649463d4b76c284b720bb7f9511f56e95deeef561fa"} Dec 11 09:15:01 crc kubenswrapper[4629]: I1211 09:15:01.456214 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" event={"ID":"b51b84fa-52d1-41b1-937b-4a5a568fd917","Type":"ContainerStarted","Data":"5ee2db38283271290f63b4e1143923db86143494ef1941322bcad8191464b725"} Dec 11 09:15:01 crc kubenswrapper[4629]: I1211 09:15:01.486068 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" podStartSLOduration=1.486030199 podStartE2EDuration="1.486030199s" podCreationTimestamp="2025-12-11 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:15:01.478553245 +0000 UTC m=+2289.543968873" watchObservedRunningTime="2025-12-11 09:15:01.486030199 +0000 UTC m=+2289.551445807" Dec 11 09:15:02 crc kubenswrapper[4629]: I1211 09:15:02.465440 4629 generic.go:334] "Generic (PLEG): container finished" podID="b51b84fa-52d1-41b1-937b-4a5a568fd917" containerID="03e5fe5e57475704cf77b649463d4b76c284b720bb7f9511f56e95deeef561fa" exitCode=0 Dec 11 09:15:02 crc kubenswrapper[4629]: I1211 09:15:02.465555 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" event={"ID":"b51b84fa-52d1-41b1-937b-4a5a568fd917","Type":"ContainerDied","Data":"03e5fe5e57475704cf77b649463d4b76c284b720bb7f9511f56e95deeef561fa"} Dec 11 09:15:03 crc kubenswrapper[4629]: I1211 09:15:03.791795 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:03 crc kubenswrapper[4629]: I1211 09:15:03.869329 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b51b84fa-52d1-41b1-937b-4a5a568fd917-secret-volume\") pod \"b51b84fa-52d1-41b1-937b-4a5a568fd917\" (UID: \"b51b84fa-52d1-41b1-937b-4a5a568fd917\") " Dec 11 09:15:03 crc kubenswrapper[4629]: I1211 09:15:03.869424 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b51b84fa-52d1-41b1-937b-4a5a568fd917-config-volume\") pod \"b51b84fa-52d1-41b1-937b-4a5a568fd917\" (UID: \"b51b84fa-52d1-41b1-937b-4a5a568fd917\") " Dec 11 09:15:03 crc kubenswrapper[4629]: I1211 09:15:03.869612 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbkqz\" (UniqueName: \"kubernetes.io/projected/b51b84fa-52d1-41b1-937b-4a5a568fd917-kube-api-access-qbkqz\") pod \"b51b84fa-52d1-41b1-937b-4a5a568fd917\" (UID: \"b51b84fa-52d1-41b1-937b-4a5a568fd917\") " Dec 11 09:15:03 crc kubenswrapper[4629]: I1211 09:15:03.870276 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b51b84fa-52d1-41b1-937b-4a5a568fd917-config-volume" (OuterVolumeSpecName: "config-volume") pod "b51b84fa-52d1-41b1-937b-4a5a568fd917" (UID: "b51b84fa-52d1-41b1-937b-4a5a568fd917"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:15:03 crc kubenswrapper[4629]: I1211 09:15:03.877987 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51b84fa-52d1-41b1-937b-4a5a568fd917-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b51b84fa-52d1-41b1-937b-4a5a568fd917" (UID: "b51b84fa-52d1-41b1-937b-4a5a568fd917"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:15:03 crc kubenswrapper[4629]: I1211 09:15:03.878050 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b51b84fa-52d1-41b1-937b-4a5a568fd917-kube-api-access-qbkqz" (OuterVolumeSpecName: "kube-api-access-qbkqz") pod "b51b84fa-52d1-41b1-937b-4a5a568fd917" (UID: "b51b84fa-52d1-41b1-937b-4a5a568fd917"). InnerVolumeSpecName "kube-api-access-qbkqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:15:03 crc kubenswrapper[4629]: I1211 09:15:03.972041 4629 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b51b84fa-52d1-41b1-937b-4a5a568fd917-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 09:15:03 crc kubenswrapper[4629]: I1211 09:15:03.972129 4629 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b51b84fa-52d1-41b1-937b-4a5a568fd917-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 09:15:03 crc kubenswrapper[4629]: I1211 09:15:03.972141 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbkqz\" (UniqueName: \"kubernetes.io/projected/b51b84fa-52d1-41b1-937b-4a5a568fd917-kube-api-access-qbkqz\") on node \"crc\" DevicePath \"\"" Dec 11 09:15:04 crc kubenswrapper[4629]: I1211 09:15:04.483299 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" event={"ID":"b51b84fa-52d1-41b1-937b-4a5a568fd917","Type":"ContainerDied","Data":"5ee2db38283271290f63b4e1143923db86143494ef1941322bcad8191464b725"} Dec 11 09:15:04 crc kubenswrapper[4629]: I1211 09:15:04.483747 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ee2db38283271290f63b4e1143923db86143494ef1941322bcad8191464b725" Dec 11 09:15:04 crc kubenswrapper[4629]: I1211 09:15:04.483368 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-mtfjf" Dec 11 09:15:04 crc kubenswrapper[4629]: I1211 09:15:04.873973 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm"] Dec 11 09:15:04 crc kubenswrapper[4629]: I1211 09:15:04.882437 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424030-vtnxm"] Dec 11 09:15:06 crc kubenswrapper[4629]: I1211 09:15:06.212829 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99df09c3-eae2-4db8-a82e-a7e1d1c462e5" path="/var/lib/kubelet/pods/99df09c3-eae2-4db8-a82e-a7e1d1c462e5/volumes" Dec 11 09:15:06 crc kubenswrapper[4629]: I1211 09:15:06.213032 4629 scope.go:117] "RemoveContainer" containerID="0b6c51b698e52977d4507e1d28e74bed00792b9aef1f5146d036958ad18dc369" Dec 11 09:15:20 crc kubenswrapper[4629]: I1211 09:15:20.486198 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:15:20 crc kubenswrapper[4629]: I1211 09:15:20.486859 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:15:50 crc kubenswrapper[4629]: I1211 09:15:50.485701 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:15:50 crc kubenswrapper[4629]: I1211 09:15:50.486351 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:15:50 crc kubenswrapper[4629]: I1211 09:15:50.486432 4629 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 09:15:50 crc kubenswrapper[4629]: I1211 09:15:50.487260 4629 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48"} pod="openshift-machine-config-operator/machine-config-daemon-wbld9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 09:15:50 crc kubenswrapper[4629]: I1211 09:15:50.487333 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" containerID="cri-o://76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" gracePeriod=600 Dec 11 09:15:50 crc kubenswrapper[4629]: E1211 09:15:50.614321 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:15:50 crc kubenswrapper[4629]: I1211 09:15:50.876747 4629 generic.go:334] "Generic (PLEG): container finished" podID="847b50b9-0093-404f-b0dd-63c814d8986b" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" exitCode=0 Dec 11 09:15:50 crc kubenswrapper[4629]: I1211 09:15:50.877174 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerDied","Data":"76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48"} Dec 11 09:15:50 crc kubenswrapper[4629]: I1211 09:15:50.877209 4629 scope.go:117] "RemoveContainer" containerID="3070d82dc8c3e310ce3a87a010814089bcbb734e3b66f41c4a60ff1e08bb1f40" Dec 11 09:15:50 crc kubenswrapper[4629]: I1211 09:15:50.877862 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:15:50 crc kubenswrapper[4629]: E1211 09:15:50.878096 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:16:05 crc kubenswrapper[4629]: I1211 09:16:05.199508 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:16:05 crc kubenswrapper[4629]: E1211 09:16:05.200273 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:16:18 crc kubenswrapper[4629]: I1211 09:16:18.198787 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:16:18 crc kubenswrapper[4629]: E1211 09:16:18.199546 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:16:33 crc kubenswrapper[4629]: I1211 09:16:33.200141 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:16:33 crc kubenswrapper[4629]: E1211 09:16:33.201521 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:16:45 crc kubenswrapper[4629]: I1211 09:16:45.198921 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:16:45 crc kubenswrapper[4629]: E1211 09:16:45.199817 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:17:00 crc kubenswrapper[4629]: I1211 09:17:00.199155 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:17:00 crc kubenswrapper[4629]: E1211 09:17:00.200063 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:17:11 crc kubenswrapper[4629]: I1211 09:17:11.199615 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:17:11 crc kubenswrapper[4629]: E1211 09:17:11.200440 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:17:24 crc kubenswrapper[4629]: I1211 09:17:24.199420 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:17:24 crc kubenswrapper[4629]: E1211 09:17:24.200380 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:17:35 crc kubenswrapper[4629]: I1211 09:17:35.199359 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:17:35 crc kubenswrapper[4629]: E1211 09:17:35.201810 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:17:48 crc kubenswrapper[4629]: I1211 09:17:48.199469 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:17:48 crc kubenswrapper[4629]: E1211 09:17:48.200546 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:18:03 crc kubenswrapper[4629]: I1211 09:18:03.199152 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:18:03 crc kubenswrapper[4629]: E1211 09:18:03.199918 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:18:16 crc kubenswrapper[4629]: I1211 09:18:16.199244 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:18:16 crc kubenswrapper[4629]: E1211 09:18:16.200101 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:18:27 crc kubenswrapper[4629]: I1211 09:18:27.198918 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:18:27 crc kubenswrapper[4629]: E1211 09:18:27.199720 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:18:41 crc kubenswrapper[4629]: I1211 09:18:41.199032 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:18:41 crc kubenswrapper[4629]: E1211 09:18:41.199950 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:18:52 crc kubenswrapper[4629]: I1211 09:18:52.205965 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:18:52 crc kubenswrapper[4629]: E1211 09:18:52.206752 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:19:07 crc kubenswrapper[4629]: I1211 09:19:07.199380 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:19:07 crc kubenswrapper[4629]: E1211 09:19:07.200374 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:19:22 crc kubenswrapper[4629]: I1211 09:19:22.205457 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:19:22 crc kubenswrapper[4629]: E1211 09:19:22.206390 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:19:37 crc kubenswrapper[4629]: I1211 09:19:37.199034 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:19:37 crc kubenswrapper[4629]: E1211 09:19:37.199983 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:19:48 crc kubenswrapper[4629]: I1211 09:19:48.199298 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:19:48 crc kubenswrapper[4629]: E1211 09:19:48.200035 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:20:01 crc kubenswrapper[4629]: I1211 09:20:01.199074 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:20:01 crc kubenswrapper[4629]: E1211 09:20:01.200918 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:20:14 crc kubenswrapper[4629]: I1211 09:20:14.199607 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:20:14 crc kubenswrapper[4629]: E1211 09:20:14.200546 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:20:28 crc kubenswrapper[4629]: I1211 09:20:28.198932 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:20:28 crc kubenswrapper[4629]: E1211 09:20:28.199710 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:20:42 crc kubenswrapper[4629]: I1211 09:20:42.207687 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:20:42 crc kubenswrapper[4629]: E1211 09:20:42.209355 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:20:53 crc kubenswrapper[4629]: I1211 09:20:53.199370 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:20:53 crc kubenswrapper[4629]: I1211 09:20:53.707125 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"d1441cf6cef81d3d1e6dde6fc7c192a9564615532b2a9036b19516be40e634b7"} Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.221906 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-65xcc/must-gather-75vsg"] Dec 11 09:22:25 crc kubenswrapper[4629]: E1211 09:22:25.222987 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51b84fa-52d1-41b1-937b-4a5a568fd917" containerName="collect-profiles" Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.223004 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51b84fa-52d1-41b1-937b-4a5a568fd917" containerName="collect-profiles" Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.223243 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51b84fa-52d1-41b1-937b-4a5a568fd917" containerName="collect-profiles" Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.224516 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-65xcc/must-gather-75vsg" Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.241057 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-65xcc"/"openshift-service-ca.crt" Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.241199 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-65xcc"/"kube-root-ca.crt" Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.252393 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5zqj\" (UniqueName: \"kubernetes.io/projected/52f7bcbe-f979-4d26-bb6b-39aa9f251c15-kube-api-access-j5zqj\") pod \"must-gather-75vsg\" (UID: \"52f7bcbe-f979-4d26-bb6b-39aa9f251c15\") " pod="openshift-must-gather-65xcc/must-gather-75vsg" Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.252784 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/52f7bcbe-f979-4d26-bb6b-39aa9f251c15-must-gather-output\") pod \"must-gather-75vsg\" (UID: \"52f7bcbe-f979-4d26-bb6b-39aa9f251c15\") " pod="openshift-must-gather-65xcc/must-gather-75vsg" Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.310758 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-65xcc/must-gather-75vsg"] Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.353989 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5zqj\" (UniqueName: \"kubernetes.io/projected/52f7bcbe-f979-4d26-bb6b-39aa9f251c15-kube-api-access-j5zqj\") pod \"must-gather-75vsg\" (UID: \"52f7bcbe-f979-4d26-bb6b-39aa9f251c15\") " pod="openshift-must-gather-65xcc/must-gather-75vsg" Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.354144 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/52f7bcbe-f979-4d26-bb6b-39aa9f251c15-must-gather-output\") pod \"must-gather-75vsg\" (UID: \"52f7bcbe-f979-4d26-bb6b-39aa9f251c15\") " pod="openshift-must-gather-65xcc/must-gather-75vsg" Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.354679 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/52f7bcbe-f979-4d26-bb6b-39aa9f251c15-must-gather-output\") pod \"must-gather-75vsg\" (UID: \"52f7bcbe-f979-4d26-bb6b-39aa9f251c15\") " pod="openshift-must-gather-65xcc/must-gather-75vsg" Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.414922 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5zqj\" (UniqueName: \"kubernetes.io/projected/52f7bcbe-f979-4d26-bb6b-39aa9f251c15-kube-api-access-j5zqj\") pod \"must-gather-75vsg\" (UID: \"52f7bcbe-f979-4d26-bb6b-39aa9f251c15\") " pod="openshift-must-gather-65xcc/must-gather-75vsg" Dec 11 09:22:25 crc kubenswrapper[4629]: I1211 09:22:25.548099 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-65xcc/must-gather-75vsg" Dec 11 09:22:26 crc kubenswrapper[4629]: I1211 09:22:26.208801 4629 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 09:22:26 crc kubenswrapper[4629]: I1211 09:22:26.218514 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-65xcc/must-gather-75vsg"] Dec 11 09:22:26 crc kubenswrapper[4629]: I1211 09:22:26.516045 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-65xcc/must-gather-75vsg" event={"ID":"52f7bcbe-f979-4d26-bb6b-39aa9f251c15","Type":"ContainerStarted","Data":"3c888efd33877530f99c9078aff47ee797d89795b656c8169d7dbf375ae20e07"} Dec 11 09:22:38 crc kubenswrapper[4629]: I1211 09:22:38.680245 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-65xcc/must-gather-75vsg" event={"ID":"52f7bcbe-f979-4d26-bb6b-39aa9f251c15","Type":"ContainerStarted","Data":"e8a4a8a01ca84c5d1a064014b7f0b6af4a8522252d5be99393ebb21fb3b57b81"} Dec 11 09:22:38 crc kubenswrapper[4629]: I1211 09:22:38.680826 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-65xcc/must-gather-75vsg" event={"ID":"52f7bcbe-f979-4d26-bb6b-39aa9f251c15","Type":"ContainerStarted","Data":"1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8"} Dec 11 09:22:38 crc kubenswrapper[4629]: I1211 09:22:38.749717 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-65xcc/must-gather-75vsg" podStartSLOduration=2.392955346 podStartE2EDuration="13.749685881s" podCreationTimestamp="2025-12-11 09:22:25 +0000 UTC" firstStartedPulling="2025-12-11 09:22:26.208725871 +0000 UTC m=+2734.274141479" lastFinishedPulling="2025-12-11 09:22:37.565456406 +0000 UTC m=+2745.630872014" observedRunningTime="2025-12-11 09:22:38.71014336 +0000 UTC m=+2746.775558978" watchObservedRunningTime="2025-12-11 09:22:38.749685881 +0000 UTC m=+2746.815101489" Dec 11 09:22:42 crc kubenswrapper[4629]: I1211 09:22:42.971460 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-65xcc/crc-debug-szd44"] Dec 11 09:22:42 crc kubenswrapper[4629]: I1211 09:22:42.973600 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-65xcc/crc-debug-szd44" Dec 11 09:22:42 crc kubenswrapper[4629]: I1211 09:22:42.975995 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-65xcc"/"default-dockercfg-5p4zf" Dec 11 09:22:43 crc kubenswrapper[4629]: I1211 09:22:43.060839 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7rzx\" (UniqueName: \"kubernetes.io/projected/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8-kube-api-access-v7rzx\") pod \"crc-debug-szd44\" (UID: \"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8\") " pod="openshift-must-gather-65xcc/crc-debug-szd44" Dec 11 09:22:43 crc kubenswrapper[4629]: I1211 09:22:43.061129 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8-host\") pod \"crc-debug-szd44\" (UID: \"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8\") " pod="openshift-must-gather-65xcc/crc-debug-szd44" Dec 11 09:22:43 crc kubenswrapper[4629]: I1211 09:22:43.162322 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8-host\") pod \"crc-debug-szd44\" (UID: \"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8\") " pod="openshift-must-gather-65xcc/crc-debug-szd44" Dec 11 09:22:43 crc kubenswrapper[4629]: I1211 09:22:43.162455 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8-host\") pod \"crc-debug-szd44\" (UID: \"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8\") " pod="openshift-must-gather-65xcc/crc-debug-szd44" Dec 11 09:22:43 crc kubenswrapper[4629]: I1211 09:22:43.162909 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7rzx\" (UniqueName: \"kubernetes.io/projected/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8-kube-api-access-v7rzx\") pod \"crc-debug-szd44\" (UID: \"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8\") " pod="openshift-must-gather-65xcc/crc-debug-szd44" Dec 11 09:22:43 crc kubenswrapper[4629]: I1211 09:22:43.191612 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7rzx\" (UniqueName: \"kubernetes.io/projected/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8-kube-api-access-v7rzx\") pod \"crc-debug-szd44\" (UID: \"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8\") " pod="openshift-must-gather-65xcc/crc-debug-szd44" Dec 11 09:22:43 crc kubenswrapper[4629]: I1211 09:22:43.298014 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-65xcc/crc-debug-szd44" Dec 11 09:22:43 crc kubenswrapper[4629]: W1211 09:22:43.338245 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73b6c6b1_f8d9_49c4_8433_9a99dde1acf8.slice/crio-4f4d1a61c76caa4b461fb6cd470243d58452a8bb5983bca84f489f14e5389154 WatchSource:0}: Error finding container 4f4d1a61c76caa4b461fb6cd470243d58452a8bb5983bca84f489f14e5389154: Status 404 returned error can't find the container with id 4f4d1a61c76caa4b461fb6cd470243d58452a8bb5983bca84f489f14e5389154 Dec 11 09:22:43 crc kubenswrapper[4629]: I1211 09:22:43.744049 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-65xcc/crc-debug-szd44" event={"ID":"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8","Type":"ContainerStarted","Data":"4f4d1a61c76caa4b461fb6cd470243d58452a8bb5983bca84f489f14e5389154"} Dec 11 09:22:54 crc kubenswrapper[4629]: I1211 09:22:54.879126 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-65xcc/crc-debug-szd44" event={"ID":"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8","Type":"ContainerStarted","Data":"91b2ed4bfbeffd04ea31c48c3c32e3a7db22a693f1cb82d0dec5abeacd7f825d"} Dec 11 09:22:54 crc kubenswrapper[4629]: I1211 09:22:54.962022 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-65xcc/crc-debug-szd44" podStartSLOduration=1.810373903 podStartE2EDuration="12.962002681s" podCreationTimestamp="2025-12-11 09:22:42 +0000 UTC" firstStartedPulling="2025-12-11 09:22:43.340112241 +0000 UTC m=+2751.405527849" lastFinishedPulling="2025-12-11 09:22:54.491741019 +0000 UTC m=+2762.557156627" observedRunningTime="2025-12-11 09:22:54.961633659 +0000 UTC m=+2763.027049287" watchObservedRunningTime="2025-12-11 09:22:54.962002681 +0000 UTC m=+2763.027418289" Dec 11 09:23:19 crc kubenswrapper[4629]: I1211 09:23:19.136144 4629 generic.go:334] "Generic (PLEG): container finished" podID="73b6c6b1-f8d9-49c4-8433-9a99dde1acf8" containerID="91b2ed4bfbeffd04ea31c48c3c32e3a7db22a693f1cb82d0dec5abeacd7f825d" exitCode=0 Dec 11 09:23:19 crc kubenswrapper[4629]: I1211 09:23:19.136309 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-65xcc/crc-debug-szd44" event={"ID":"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8","Type":"ContainerDied","Data":"91b2ed4bfbeffd04ea31c48c3c32e3a7db22a693f1cb82d0dec5abeacd7f825d"} Dec 11 09:23:20 crc kubenswrapper[4629]: I1211 09:23:20.407619 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-65xcc/crc-debug-szd44" Dec 11 09:23:20 crc kubenswrapper[4629]: I1211 09:23:20.458791 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-65xcc/crc-debug-szd44"] Dec 11 09:23:20 crc kubenswrapper[4629]: I1211 09:23:20.470229 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-65xcc/crc-debug-szd44"] Dec 11 09:23:20 crc kubenswrapper[4629]: I1211 09:23:20.485640 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:23:20 crc kubenswrapper[4629]: I1211 09:23:20.485727 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:23:20 crc kubenswrapper[4629]: I1211 09:23:20.532048 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7rzx\" (UniqueName: \"kubernetes.io/projected/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8-kube-api-access-v7rzx\") pod \"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8\" (UID: \"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8\") " Dec 11 09:23:20 crc kubenswrapper[4629]: I1211 09:23:20.532184 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8-host\") pod \"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8\" (UID: \"73b6c6b1-f8d9-49c4-8433-9a99dde1acf8\") " Dec 11 09:23:20 crc kubenswrapper[4629]: I1211 09:23:20.532621 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8-host" (OuterVolumeSpecName: "host") pod "73b6c6b1-f8d9-49c4-8433-9a99dde1acf8" (UID: "73b6c6b1-f8d9-49c4-8433-9a99dde1acf8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 09:23:20 crc kubenswrapper[4629]: I1211 09:23:20.551052 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8-kube-api-access-v7rzx" (OuterVolumeSpecName: "kube-api-access-v7rzx") pod "73b6c6b1-f8d9-49c4-8433-9a99dde1acf8" (UID: "73b6c6b1-f8d9-49c4-8433-9a99dde1acf8"). InnerVolumeSpecName "kube-api-access-v7rzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:23:20 crc kubenswrapper[4629]: I1211 09:23:20.634193 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7rzx\" (UniqueName: \"kubernetes.io/projected/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8-kube-api-access-v7rzx\") on node \"crc\" DevicePath \"\"" Dec 11 09:23:20 crc kubenswrapper[4629]: I1211 09:23:20.634451 4629 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8-host\") on node \"crc\" DevicePath \"\"" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.167055 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f4d1a61c76caa4b461fb6cd470243d58452a8bb5983bca84f489f14e5389154" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.167142 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-65xcc/crc-debug-szd44" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.631822 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-65xcc/crc-debug-2qwwd"] Dec 11 09:23:21 crc kubenswrapper[4629]: E1211 09:23:21.632202 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b6c6b1-f8d9-49c4-8433-9a99dde1acf8" containerName="container-00" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.632214 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b6c6b1-f8d9-49c4-8433-9a99dde1acf8" containerName="container-00" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.632428 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="73b6c6b1-f8d9-49c4-8433-9a99dde1acf8" containerName="container-00" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.633037 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-65xcc/crc-debug-2qwwd" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.635243 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-65xcc"/"default-dockercfg-5p4zf" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.755326 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsr2g\" (UniqueName: \"kubernetes.io/projected/f176c7f7-e348-4531-9680-e47b88968ddf-kube-api-access-lsr2g\") pod \"crc-debug-2qwwd\" (UID: \"f176c7f7-e348-4531-9680-e47b88968ddf\") " pod="openshift-must-gather-65xcc/crc-debug-2qwwd" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.755382 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f176c7f7-e348-4531-9680-e47b88968ddf-host\") pod \"crc-debug-2qwwd\" (UID: \"f176c7f7-e348-4531-9680-e47b88968ddf\") " pod="openshift-must-gather-65xcc/crc-debug-2qwwd" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.856777 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsr2g\" (UniqueName: \"kubernetes.io/projected/f176c7f7-e348-4531-9680-e47b88968ddf-kube-api-access-lsr2g\") pod \"crc-debug-2qwwd\" (UID: \"f176c7f7-e348-4531-9680-e47b88968ddf\") " pod="openshift-must-gather-65xcc/crc-debug-2qwwd" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.856865 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f176c7f7-e348-4531-9680-e47b88968ddf-host\") pod \"crc-debug-2qwwd\" (UID: \"f176c7f7-e348-4531-9680-e47b88968ddf\") " pod="openshift-must-gather-65xcc/crc-debug-2qwwd" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.856991 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f176c7f7-e348-4531-9680-e47b88968ddf-host\") pod \"crc-debug-2qwwd\" (UID: \"f176c7f7-e348-4531-9680-e47b88968ddf\") " pod="openshift-must-gather-65xcc/crc-debug-2qwwd" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.875327 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsr2g\" (UniqueName: \"kubernetes.io/projected/f176c7f7-e348-4531-9680-e47b88968ddf-kube-api-access-lsr2g\") pod \"crc-debug-2qwwd\" (UID: \"f176c7f7-e348-4531-9680-e47b88968ddf\") " pod="openshift-must-gather-65xcc/crc-debug-2qwwd" Dec 11 09:23:21 crc kubenswrapper[4629]: I1211 09:23:21.951411 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-65xcc/crc-debug-2qwwd" Dec 11 09:23:22 crc kubenswrapper[4629]: I1211 09:23:22.177172 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-65xcc/crc-debug-2qwwd" event={"ID":"f176c7f7-e348-4531-9680-e47b88968ddf","Type":"ContainerStarted","Data":"8f04b4f88befe1bf72be86fa2d89f9e909ed2db617ebb475e35e5df1b663c72b"} Dec 11 09:23:22 crc kubenswrapper[4629]: I1211 09:23:22.237219 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73b6c6b1-f8d9-49c4-8433-9a99dde1acf8" path="/var/lib/kubelet/pods/73b6c6b1-f8d9-49c4-8433-9a99dde1acf8/volumes" Dec 11 09:23:23 crc kubenswrapper[4629]: I1211 09:23:23.187749 4629 generic.go:334] "Generic (PLEG): container finished" podID="f176c7f7-e348-4531-9680-e47b88968ddf" containerID="1d833edd362ab615fdb2ba73ba31b8b574b18098725b5e7989c217517fdf9f7e" exitCode=1 Dec 11 09:23:23 crc kubenswrapper[4629]: I1211 09:23:23.187863 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-65xcc/crc-debug-2qwwd" event={"ID":"f176c7f7-e348-4531-9680-e47b88968ddf","Type":"ContainerDied","Data":"1d833edd362ab615fdb2ba73ba31b8b574b18098725b5e7989c217517fdf9f7e"} Dec 11 09:23:23 crc kubenswrapper[4629]: I1211 09:23:23.239601 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-65xcc/crc-debug-2qwwd"] Dec 11 09:23:23 crc kubenswrapper[4629]: I1211 09:23:23.249249 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-65xcc/crc-debug-2qwwd"] Dec 11 09:23:24 crc kubenswrapper[4629]: I1211 09:23:24.308966 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-65xcc/crc-debug-2qwwd" Dec 11 09:23:24 crc kubenswrapper[4629]: I1211 09:23:24.404748 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f176c7f7-e348-4531-9680-e47b88968ddf-host\") pod \"f176c7f7-e348-4531-9680-e47b88968ddf\" (UID: \"f176c7f7-e348-4531-9680-e47b88968ddf\") " Dec 11 09:23:24 crc kubenswrapper[4629]: I1211 09:23:24.404939 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f176c7f7-e348-4531-9680-e47b88968ddf-host" (OuterVolumeSpecName: "host") pod "f176c7f7-e348-4531-9680-e47b88968ddf" (UID: "f176c7f7-e348-4531-9680-e47b88968ddf"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 09:23:24 crc kubenswrapper[4629]: I1211 09:23:24.405048 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsr2g\" (UniqueName: \"kubernetes.io/projected/f176c7f7-e348-4531-9680-e47b88968ddf-kube-api-access-lsr2g\") pod \"f176c7f7-e348-4531-9680-e47b88968ddf\" (UID: \"f176c7f7-e348-4531-9680-e47b88968ddf\") " Dec 11 09:23:24 crc kubenswrapper[4629]: I1211 09:23:24.405567 4629 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f176c7f7-e348-4531-9680-e47b88968ddf-host\") on node \"crc\" DevicePath \"\"" Dec 11 09:23:24 crc kubenswrapper[4629]: I1211 09:23:24.418080 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f176c7f7-e348-4531-9680-e47b88968ddf-kube-api-access-lsr2g" (OuterVolumeSpecName: "kube-api-access-lsr2g") pod "f176c7f7-e348-4531-9680-e47b88968ddf" (UID: "f176c7f7-e348-4531-9680-e47b88968ddf"). InnerVolumeSpecName "kube-api-access-lsr2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:23:24 crc kubenswrapper[4629]: I1211 09:23:24.507239 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsr2g\" (UniqueName: \"kubernetes.io/projected/f176c7f7-e348-4531-9680-e47b88968ddf-kube-api-access-lsr2g\") on node \"crc\" DevicePath \"\"" Dec 11 09:23:25 crc kubenswrapper[4629]: I1211 09:23:25.206367 4629 scope.go:117] "RemoveContainer" containerID="1d833edd362ab615fdb2ba73ba31b8b574b18098725b5e7989c217517fdf9f7e" Dec 11 09:23:25 crc kubenswrapper[4629]: I1211 09:23:25.206742 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-65xcc/crc-debug-2qwwd" Dec 11 09:23:26 crc kubenswrapper[4629]: I1211 09:23:26.210765 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f176c7f7-e348-4531-9680-e47b88968ddf" path="/var/lib/kubelet/pods/f176c7f7-e348-4531-9680-e47b88968ddf/volumes" Dec 11 09:23:39 crc kubenswrapper[4629]: I1211 09:23:39.954431 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gnt6g"] Dec 11 09:23:39 crc kubenswrapper[4629]: E1211 09:23:39.965567 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f176c7f7-e348-4531-9680-e47b88968ddf" containerName="container-00" Dec 11 09:23:39 crc kubenswrapper[4629]: I1211 09:23:39.965589 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="f176c7f7-e348-4531-9680-e47b88968ddf" containerName="container-00" Dec 11 09:23:39 crc kubenswrapper[4629]: I1211 09:23:39.965761 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="f176c7f7-e348-4531-9680-e47b88968ddf" containerName="container-00" Dec 11 09:23:39 crc kubenswrapper[4629]: I1211 09:23:39.967770 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnt6g"] Dec 11 09:23:39 crc kubenswrapper[4629]: I1211 09:23:39.968230 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:39 crc kubenswrapper[4629]: I1211 09:23:39.998820 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2060d1a6-62e9-4a5e-b36c-090fb9292999-catalog-content\") pod \"redhat-marketplace-gnt6g\" (UID: \"2060d1a6-62e9-4a5e-b36c-090fb9292999\") " pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:39 crc kubenswrapper[4629]: I1211 09:23:39.999258 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp2tz\" (UniqueName: \"kubernetes.io/projected/2060d1a6-62e9-4a5e-b36c-090fb9292999-kube-api-access-cp2tz\") pod \"redhat-marketplace-gnt6g\" (UID: \"2060d1a6-62e9-4a5e-b36c-090fb9292999\") " pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:39 crc kubenswrapper[4629]: I1211 09:23:39.999407 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2060d1a6-62e9-4a5e-b36c-090fb9292999-utilities\") pod \"redhat-marketplace-gnt6g\" (UID: \"2060d1a6-62e9-4a5e-b36c-090fb9292999\") " pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:40 crc kubenswrapper[4629]: I1211 09:23:40.101222 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2060d1a6-62e9-4a5e-b36c-090fb9292999-catalog-content\") pod \"redhat-marketplace-gnt6g\" (UID: \"2060d1a6-62e9-4a5e-b36c-090fb9292999\") " pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:40 crc kubenswrapper[4629]: I1211 09:23:40.101280 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp2tz\" (UniqueName: \"kubernetes.io/projected/2060d1a6-62e9-4a5e-b36c-090fb9292999-kube-api-access-cp2tz\") pod \"redhat-marketplace-gnt6g\" (UID: \"2060d1a6-62e9-4a5e-b36c-090fb9292999\") " pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:40 crc kubenswrapper[4629]: I1211 09:23:40.101362 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2060d1a6-62e9-4a5e-b36c-090fb9292999-utilities\") pod \"redhat-marketplace-gnt6g\" (UID: \"2060d1a6-62e9-4a5e-b36c-090fb9292999\") " pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:40 crc kubenswrapper[4629]: I1211 09:23:40.101868 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2060d1a6-62e9-4a5e-b36c-090fb9292999-catalog-content\") pod \"redhat-marketplace-gnt6g\" (UID: \"2060d1a6-62e9-4a5e-b36c-090fb9292999\") " pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:40 crc kubenswrapper[4629]: I1211 09:23:40.102041 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2060d1a6-62e9-4a5e-b36c-090fb9292999-utilities\") pod \"redhat-marketplace-gnt6g\" (UID: \"2060d1a6-62e9-4a5e-b36c-090fb9292999\") " pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:40 crc kubenswrapper[4629]: I1211 09:23:40.128872 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp2tz\" (UniqueName: \"kubernetes.io/projected/2060d1a6-62e9-4a5e-b36c-090fb9292999-kube-api-access-cp2tz\") pod \"redhat-marketplace-gnt6g\" (UID: \"2060d1a6-62e9-4a5e-b36c-090fb9292999\") " pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:40 crc kubenswrapper[4629]: I1211 09:23:40.290114 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:40 crc kubenswrapper[4629]: I1211 09:23:40.786700 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnt6g"] Dec 11 09:23:41 crc kubenswrapper[4629]: I1211 09:23:41.365961 4629 generic.go:334] "Generic (PLEG): container finished" podID="2060d1a6-62e9-4a5e-b36c-090fb9292999" containerID="29dc1de4f8323d4b371fb37716931017c09b99c372b2eb4dcf487738e0629ed7" exitCode=0 Dec 11 09:23:41 crc kubenswrapper[4629]: I1211 09:23:41.366243 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnt6g" event={"ID":"2060d1a6-62e9-4a5e-b36c-090fb9292999","Type":"ContainerDied","Data":"29dc1de4f8323d4b371fb37716931017c09b99c372b2eb4dcf487738e0629ed7"} Dec 11 09:23:41 crc kubenswrapper[4629]: I1211 09:23:41.366316 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnt6g" event={"ID":"2060d1a6-62e9-4a5e-b36c-090fb9292999","Type":"ContainerStarted","Data":"61a4c66fbcf9ce97a7071df93cae5ceb7a55c2a0cb12f61162daffeb2d1a6aac"} Dec 11 09:23:43 crc kubenswrapper[4629]: I1211 09:23:43.387110 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnt6g" event={"ID":"2060d1a6-62e9-4a5e-b36c-090fb9292999","Type":"ContainerStarted","Data":"c75eae6054ffb3d5faa6700a0b2eae741078a68adbe58f019e57aa6c303f4b92"} Dec 11 09:23:44 crc kubenswrapper[4629]: I1211 09:23:44.397188 4629 generic.go:334] "Generic (PLEG): container finished" podID="2060d1a6-62e9-4a5e-b36c-090fb9292999" containerID="c75eae6054ffb3d5faa6700a0b2eae741078a68adbe58f019e57aa6c303f4b92" exitCode=0 Dec 11 09:23:44 crc kubenswrapper[4629]: I1211 09:23:44.397318 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnt6g" event={"ID":"2060d1a6-62e9-4a5e-b36c-090fb9292999","Type":"ContainerDied","Data":"c75eae6054ffb3d5faa6700a0b2eae741078a68adbe58f019e57aa6c303f4b92"} Dec 11 09:23:46 crc kubenswrapper[4629]: I1211 09:23:46.415283 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnt6g" event={"ID":"2060d1a6-62e9-4a5e-b36c-090fb9292999","Type":"ContainerStarted","Data":"0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2"} Dec 11 09:23:46 crc kubenswrapper[4629]: I1211 09:23:46.440973 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gnt6g" podStartSLOduration=3.230194963 podStartE2EDuration="7.440957085s" podCreationTimestamp="2025-12-11 09:23:39 +0000 UTC" firstStartedPulling="2025-12-11 09:23:41.368974717 +0000 UTC m=+2809.434390325" lastFinishedPulling="2025-12-11 09:23:45.579736839 +0000 UTC m=+2813.645152447" observedRunningTime="2025-12-11 09:23:46.436203165 +0000 UTC m=+2814.501618773" watchObservedRunningTime="2025-12-11 09:23:46.440957085 +0000 UTC m=+2814.506372693" Dec 11 09:23:50 crc kubenswrapper[4629]: I1211 09:23:50.290813 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:50 crc kubenswrapper[4629]: I1211 09:23:50.291466 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:50 crc kubenswrapper[4629]: I1211 09:23:50.338573 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:50 crc kubenswrapper[4629]: I1211 09:23:50.486439 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:23:50 crc kubenswrapper[4629]: I1211 09:23:50.486634 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:23:50 crc kubenswrapper[4629]: I1211 09:23:50.502257 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:50 crc kubenswrapper[4629]: I1211 09:23:50.593416 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnt6g"] Dec 11 09:23:52 crc kubenswrapper[4629]: I1211 09:23:52.461801 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gnt6g" podUID="2060d1a6-62e9-4a5e-b36c-090fb9292999" containerName="registry-server" containerID="cri-o://0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2" gracePeriod=2 Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.019495 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.041470 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2060d1a6-62e9-4a5e-b36c-090fb9292999-catalog-content\") pod \"2060d1a6-62e9-4a5e-b36c-090fb9292999\" (UID: \"2060d1a6-62e9-4a5e-b36c-090fb9292999\") " Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.041605 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2060d1a6-62e9-4a5e-b36c-090fb9292999-utilities\") pod \"2060d1a6-62e9-4a5e-b36c-090fb9292999\" (UID: \"2060d1a6-62e9-4a5e-b36c-090fb9292999\") " Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.041725 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp2tz\" (UniqueName: \"kubernetes.io/projected/2060d1a6-62e9-4a5e-b36c-090fb9292999-kube-api-access-cp2tz\") pod \"2060d1a6-62e9-4a5e-b36c-090fb9292999\" (UID: \"2060d1a6-62e9-4a5e-b36c-090fb9292999\") " Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.042966 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2060d1a6-62e9-4a5e-b36c-090fb9292999-utilities" (OuterVolumeSpecName: "utilities") pod "2060d1a6-62e9-4a5e-b36c-090fb9292999" (UID: "2060d1a6-62e9-4a5e-b36c-090fb9292999"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.048190 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2060d1a6-62e9-4a5e-b36c-090fb9292999-kube-api-access-cp2tz" (OuterVolumeSpecName: "kube-api-access-cp2tz") pod "2060d1a6-62e9-4a5e-b36c-090fb9292999" (UID: "2060d1a6-62e9-4a5e-b36c-090fb9292999"). InnerVolumeSpecName "kube-api-access-cp2tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.138450 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2060d1a6-62e9-4a5e-b36c-090fb9292999-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2060d1a6-62e9-4a5e-b36c-090fb9292999" (UID: "2060d1a6-62e9-4a5e-b36c-090fb9292999"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.153959 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp2tz\" (UniqueName: \"kubernetes.io/projected/2060d1a6-62e9-4a5e-b36c-090fb9292999-kube-api-access-cp2tz\") on node \"crc\" DevicePath \"\"" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.154002 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2060d1a6-62e9-4a5e-b36c-090fb9292999-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.154012 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2060d1a6-62e9-4a5e-b36c-090fb9292999-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.471020 4629 generic.go:334] "Generic (PLEG): container finished" podID="2060d1a6-62e9-4a5e-b36c-090fb9292999" containerID="0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2" exitCode=0 Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.471063 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnt6g" event={"ID":"2060d1a6-62e9-4a5e-b36c-090fb9292999","Type":"ContainerDied","Data":"0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2"} Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.471093 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnt6g" event={"ID":"2060d1a6-62e9-4a5e-b36c-090fb9292999","Type":"ContainerDied","Data":"61a4c66fbcf9ce97a7071df93cae5ceb7a55c2a0cb12f61162daffeb2d1a6aac"} Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.471110 4629 scope.go:117] "RemoveContainer" containerID="0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.471232 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnt6g" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.499341 4629 scope.go:117] "RemoveContainer" containerID="c75eae6054ffb3d5faa6700a0b2eae741078a68adbe58f019e57aa6c303f4b92" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.511590 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnt6g"] Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.528138 4629 scope.go:117] "RemoveContainer" containerID="29dc1de4f8323d4b371fb37716931017c09b99c372b2eb4dcf487738e0629ed7" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.532726 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnt6g"] Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.573357 4629 scope.go:117] "RemoveContainer" containerID="0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2" Dec 11 09:23:53 crc kubenswrapper[4629]: E1211 09:23:53.573988 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2\": container with ID starting with 0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2 not found: ID does not exist" containerID="0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.574044 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2"} err="failed to get container status \"0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2\": rpc error: code = NotFound desc = could not find container \"0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2\": container with ID starting with 0edacd81d3da91b899ac94276ef55a89b721ebb2ae17893c0866170a95d3a1a2 not found: ID does not exist" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.574077 4629 scope.go:117] "RemoveContainer" containerID="c75eae6054ffb3d5faa6700a0b2eae741078a68adbe58f019e57aa6c303f4b92" Dec 11 09:23:53 crc kubenswrapper[4629]: E1211 09:23:53.574434 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c75eae6054ffb3d5faa6700a0b2eae741078a68adbe58f019e57aa6c303f4b92\": container with ID starting with c75eae6054ffb3d5faa6700a0b2eae741078a68adbe58f019e57aa6c303f4b92 not found: ID does not exist" containerID="c75eae6054ffb3d5faa6700a0b2eae741078a68adbe58f019e57aa6c303f4b92" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.574469 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c75eae6054ffb3d5faa6700a0b2eae741078a68adbe58f019e57aa6c303f4b92"} err="failed to get container status \"c75eae6054ffb3d5faa6700a0b2eae741078a68adbe58f019e57aa6c303f4b92\": rpc error: code = NotFound desc = could not find container \"c75eae6054ffb3d5faa6700a0b2eae741078a68adbe58f019e57aa6c303f4b92\": container with ID starting with c75eae6054ffb3d5faa6700a0b2eae741078a68adbe58f019e57aa6c303f4b92 not found: ID does not exist" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.574498 4629 scope.go:117] "RemoveContainer" containerID="29dc1de4f8323d4b371fb37716931017c09b99c372b2eb4dcf487738e0629ed7" Dec 11 09:23:53 crc kubenswrapper[4629]: E1211 09:23:53.574798 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29dc1de4f8323d4b371fb37716931017c09b99c372b2eb4dcf487738e0629ed7\": container with ID starting with 29dc1de4f8323d4b371fb37716931017c09b99c372b2eb4dcf487738e0629ed7 not found: ID does not exist" containerID="29dc1de4f8323d4b371fb37716931017c09b99c372b2eb4dcf487738e0629ed7" Dec 11 09:23:53 crc kubenswrapper[4629]: I1211 09:23:53.574876 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29dc1de4f8323d4b371fb37716931017c09b99c372b2eb4dcf487738e0629ed7"} err="failed to get container status \"29dc1de4f8323d4b371fb37716931017c09b99c372b2eb4dcf487738e0629ed7\": rpc error: code = NotFound desc = could not find container \"29dc1de4f8323d4b371fb37716931017c09b99c372b2eb4dcf487738e0629ed7\": container with ID starting with 29dc1de4f8323d4b371fb37716931017c09b99c372b2eb4dcf487738e0629ed7 not found: ID does not exist" Dec 11 09:23:54 crc kubenswrapper[4629]: I1211 09:23:54.210440 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2060d1a6-62e9-4a5e-b36c-090fb9292999" path="/var/lib/kubelet/pods/2060d1a6-62e9-4a5e-b36c-090fb9292999/volumes" Dec 11 09:24:05 crc kubenswrapper[4629]: I1211 09:24:05.059817 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-56ffd55c5d-fzz6k_e730a359-2929-4370-b853-33a09bc7180c/barbican-api/0.log" Dec 11 09:24:05 crc kubenswrapper[4629]: I1211 09:24:05.183492 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-56ffd55c5d-fzz6k_e730a359-2929-4370-b853-33a09bc7180c/barbican-api-log/0.log" Dec 11 09:24:05 crc kubenswrapper[4629]: I1211 09:24:05.304535 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-695db84c66-bxp2x_d220d34a-68ce-4217-a15c-4d994b87d988/barbican-keystone-listener/0.log" Dec 11 09:24:05 crc kubenswrapper[4629]: I1211 09:24:05.333758 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-695db84c66-bxp2x_d220d34a-68ce-4217-a15c-4d994b87d988/barbican-keystone-listener-log/0.log" Dec 11 09:24:05 crc kubenswrapper[4629]: I1211 09:24:05.486954 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-84ddb5cbd7-jjz7r_3a9fa1b9-acec-4136-9819-d76fa0f88183/barbican-worker/0.log" Dec 11 09:24:05 crc kubenswrapper[4629]: I1211 09:24:05.578466 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-84ddb5cbd7-jjz7r_3a9fa1b9-acec-4136-9819-d76fa0f88183/barbican-worker-log/0.log" Dec 11 09:24:05 crc kubenswrapper[4629]: I1211 09:24:05.768287 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4jg6w_d898802e-842a-468e-a468-8aa61c2c9360/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:24:05 crc kubenswrapper[4629]: I1211 09:24:05.904045 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a9568824-b16c-4084-90c7-dd6fe60a866a/ceilometer-central-agent/0.log" Dec 11 09:24:05 crc kubenswrapper[4629]: I1211 09:24:05.967929 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a9568824-b16c-4084-90c7-dd6fe60a866a/proxy-httpd/0.log" Dec 11 09:24:05 crc kubenswrapper[4629]: I1211 09:24:05.989686 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a9568824-b16c-4084-90c7-dd6fe60a866a/ceilometer-notification-agent/0.log" Dec 11 09:24:06 crc kubenswrapper[4629]: I1211 09:24:06.016519 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a9568824-b16c-4084-90c7-dd6fe60a866a/sg-core/0.log" Dec 11 09:24:06 crc kubenswrapper[4629]: I1211 09:24:06.282172 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j54ng_c5692012-6ddd-4436-8f60-bb0ebba59d81/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:24:06 crc kubenswrapper[4629]: I1211 09:24:06.365008 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_be28a394-a248-4f19-83d0-b7408648fa5d/cinder-api/0.log" Dec 11 09:24:06 crc kubenswrapper[4629]: I1211 09:24:06.528747 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_be28a394-a248-4f19-83d0-b7408648fa5d/cinder-api-log/0.log" Dec 11 09:24:06 crc kubenswrapper[4629]: I1211 09:24:06.648484 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_648e881c-c342-458f-8acd-71f7e1bff9a0/cinder-scheduler/0.log" Dec 11 09:24:06 crc kubenswrapper[4629]: I1211 09:24:06.729964 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_648e881c-c342-458f-8acd-71f7e1bff9a0/probe/0.log" Dec 11 09:24:06 crc kubenswrapper[4629]: I1211 09:24:06.856464 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-5p5rq_ea77da00-5dbb-4f8a-883c-3c8984c55122/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:24:06 crc kubenswrapper[4629]: I1211 09:24:06.991523 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-7gkld_99f4ba2c-fa6c-438f-8c01-90663ffc9f80/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:24:07 crc kubenswrapper[4629]: I1211 09:24:07.251923 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-667ff9c869-m6rlw_12aa6e1b-5b39-48c9-80bf-6f57d8f5333e/init/0.log" Dec 11 09:24:07 crc kubenswrapper[4629]: I1211 09:24:07.530497 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-667ff9c869-m6rlw_12aa6e1b-5b39-48c9-80bf-6f57d8f5333e/dnsmasq-dns/0.log" Dec 11 09:24:07 crc kubenswrapper[4629]: I1211 09:24:07.550114 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-667ff9c869-m6rlw_12aa6e1b-5b39-48c9-80bf-6f57d8f5333e/init/0.log" Dec 11 09:24:07 crc kubenswrapper[4629]: I1211 09:24:07.558259 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-xrltk_f417ce11-881e-4571-913b-3b41106c935e/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:24:07 crc kubenswrapper[4629]: I1211 09:24:07.765179 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5fccc9499c-f4wbp_2998e2d9-9cba-4c82-ac41-9080e89798a3/keystone-api/0.log" Dec 11 09:24:07 crc kubenswrapper[4629]: I1211 09:24:07.826717 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29424061-mzwc9_cef5a8b9-b204-44a9-92fd-5bd1405895dd/keystone-cron/0.log" Dec 11 09:24:08 crc kubenswrapper[4629]: I1211 09:24:08.161838 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_005eb316-fed1-4771-9934-3d86cc6d02d3/kube-state-metrics/0.log" Dec 11 09:24:08 crc kubenswrapper[4629]: I1211 09:24:08.441923 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-59df6557fc-qgbb6_1446ae15-0cba-4d28-ae20-0b18ef295ccc/neutron-api/0.log" Dec 11 09:24:08 crc kubenswrapper[4629]: I1211 09:24:08.458743 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-59df6557fc-qgbb6_1446ae15-0cba-4d28-ae20-0b18ef295ccc/neutron-httpd/0.log" Dec 11 09:24:09 crc kubenswrapper[4629]: I1211 09:24:09.123341 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_12842a16-5854-4f02-932e-24c306f4a07b/nova-api-log/0.log" Dec 11 09:24:09 crc kubenswrapper[4629]: I1211 09:24:09.198406 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_12842a16-5854-4f02-932e-24c306f4a07b/nova-api-api/0.log" Dec 11 09:24:09 crc kubenswrapper[4629]: I1211 09:24:09.694755 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_3c34c47c-61ee-45e6-a289-5562eb5451e5/nova-cell0-conductor-conductor/0.log" Dec 11 09:24:09 crc kubenswrapper[4629]: I1211 09:24:09.715320 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_e3ea5104-e7b6-41fc-a04a-e1292739c878/nova-cell1-conductor-conductor/0.log" Dec 11 09:24:10 crc kubenswrapper[4629]: I1211 09:24:10.288523 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_82aa7b94-e1ac-4a42-a0d9-308c6c2d8cf7/nova-cell1-novncproxy-novncproxy/0.log" Dec 11 09:24:10 crc kubenswrapper[4629]: I1211 09:24:10.625610 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ff06114c-4d4f-4868-a19c-0598aee0ab52/nova-metadata-log/0.log" Dec 11 09:24:10 crc kubenswrapper[4629]: I1211 09:24:10.857231 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_7f99fda5-5b51-4936-a64f-c9e210a1fc64/nova-scheduler-scheduler/0.log" Dec 11 09:24:10 crc kubenswrapper[4629]: I1211 09:24:10.869268 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ab910fe2-4e06-4174-9638-e9b920c5bff9/mysql-bootstrap/0.log" Dec 11 09:24:11 crc kubenswrapper[4629]: I1211 09:24:11.121618 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ab910fe2-4e06-4174-9638-e9b920c5bff9/mysql-bootstrap/0.log" Dec 11 09:24:11 crc kubenswrapper[4629]: I1211 09:24:11.254103 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ab910fe2-4e06-4174-9638-e9b920c5bff9/galera/0.log" Dec 11 09:24:11 crc kubenswrapper[4629]: I1211 09:24:11.298029 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ff06114c-4d4f-4868-a19c-0598aee0ab52/nova-metadata-metadata/0.log" Dec 11 09:24:11 crc kubenswrapper[4629]: I1211 09:24:11.482502 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_762352f0-3934-4d44-b846-fa922ffedd1d/mysql-bootstrap/0.log" Dec 11 09:24:11 crc kubenswrapper[4629]: I1211 09:24:11.688264 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_762352f0-3934-4d44-b846-fa922ffedd1d/mysql-bootstrap/0.log" Dec 11 09:24:11 crc kubenswrapper[4629]: I1211 09:24:11.711246 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_762352f0-3934-4d44-b846-fa922ffedd1d/galera/0.log" Dec 11 09:24:11 crc kubenswrapper[4629]: I1211 09:24:11.779571 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_9eb2d44d-ec14-4f24-901b-ba24da452c54/openstackclient/0.log" Dec 11 09:24:11 crc kubenswrapper[4629]: I1211 09:24:11.976758 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-8lfjz_708ce743-d291-41e4-8c15-a9ad1cda240e/ovn-controller/0.log" Dec 11 09:24:12 crc kubenswrapper[4629]: I1211 09:24:12.123300 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vzhn9_8f807dbd-baed-4ee5-b422-8a3a4e2d045f/openstack-network-exporter/0.log" Dec 11 09:24:12 crc kubenswrapper[4629]: I1211 09:24:12.362019 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v9ctv_5a52781a-0193-4e3a-b0d3-4ee50d507f1a/ovsdb-server-init/0.log" Dec 11 09:24:12 crc kubenswrapper[4629]: I1211 09:24:12.639636 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v9ctv_5a52781a-0193-4e3a-b0d3-4ee50d507f1a/ovs-vswitchd/0.log" Dec 11 09:24:12 crc kubenswrapper[4629]: I1211 09:24:12.643441 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v9ctv_5a52781a-0193-4e3a-b0d3-4ee50d507f1a/ovsdb-server/0.log" Dec 11 09:24:12 crc kubenswrapper[4629]: I1211 09:24:12.674024 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v9ctv_5a52781a-0193-4e3a-b0d3-4ee50d507f1a/ovsdb-server-init/0.log" Dec 11 09:24:12 crc kubenswrapper[4629]: I1211 09:24:12.948533 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_0ccfb64c-586b-4381-96b7-5846746d6433/openstack-network-exporter/0.log" Dec 11 09:24:12 crc kubenswrapper[4629]: I1211 09:24:12.950706 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f8967f98-185a-4b7e-a43b-ff0555430bc1/openstack-network-exporter/0.log" Dec 11 09:24:12 crc kubenswrapper[4629]: I1211 09:24:12.952426 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f8967f98-185a-4b7e-a43b-ff0555430bc1/ovn-northd/0.log" Dec 11 09:24:13 crc kubenswrapper[4629]: I1211 09:24:13.194344 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_0ccfb64c-586b-4381-96b7-5846746d6433/ovsdbserver-nb/0.log" Dec 11 09:24:13 crc kubenswrapper[4629]: I1211 09:24:13.264939 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d2811a7c-5ac9-4db5-81d3-61771795a87e/openstack-network-exporter/0.log" Dec 11 09:24:13 crc kubenswrapper[4629]: I1211 09:24:13.367039 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d2811a7c-5ac9-4db5-81d3-61771795a87e/ovsdbserver-sb/0.log" Dec 11 09:24:13 crc kubenswrapper[4629]: I1211 09:24:13.576094 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-684987dfb8-72476_9d2a333e-3183-4709-a7f9-a04d60b90bad/placement-api/0.log" Dec 11 09:24:13 crc kubenswrapper[4629]: I1211 09:24:13.682595 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-684987dfb8-72476_9d2a333e-3183-4709-a7f9-a04d60b90bad/placement-log/0.log" Dec 11 09:24:13 crc kubenswrapper[4629]: I1211 09:24:13.842417 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_82957ad9-6716-4f92-beb1-dd7502ece10d/setup-container/0.log" Dec 11 09:24:14 crc kubenswrapper[4629]: I1211 09:24:14.150663 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_82957ad9-6716-4f92-beb1-dd7502ece10d/rabbitmq/0.log" Dec 11 09:24:14 crc kubenswrapper[4629]: I1211 09:24:14.164086 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_82957ad9-6716-4f92-beb1-dd7502ece10d/setup-container/0.log" Dec 11 09:24:14 crc kubenswrapper[4629]: I1211 09:24:14.266400 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8c10d7a6-61cb-4dd8-884c-1672193f2d16/setup-container/0.log" Dec 11 09:24:14 crc kubenswrapper[4629]: I1211 09:24:14.563150 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8c10d7a6-61cb-4dd8-884c-1672193f2d16/rabbitmq/0.log" Dec 11 09:24:14 crc kubenswrapper[4629]: I1211 09:24:14.574672 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8c10d7a6-61cb-4dd8-884c-1672193f2d16/setup-container/0.log" Dec 11 09:24:14 crc kubenswrapper[4629]: I1211 09:24:14.657181 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-2kcmn_636e5c8f-f6f4-4cc7-bf3b-d5da5a1eb4e0/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:24:14 crc kubenswrapper[4629]: I1211 09:24:14.813821 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vhmwf_0206ce20-df99-494c-96f0-f6201de98376/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:24:14 crc kubenswrapper[4629]: I1211 09:24:14.901543 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-lmqx2_22bb894d-6f5b-4e9e-9cca-b2d75ee4220a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:24:15 crc kubenswrapper[4629]: I1211 09:24:15.110112 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ztllp_c5239464-94d3-4812-aad7-6c864eb46731/ssh-known-hosts-edpm-deployment/0.log" Dec 11 09:24:15 crc kubenswrapper[4629]: I1211 09:24:15.336347 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-s4dbt_22acf22a-3fc4-4cf4-88e1-bbd5863c906c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:24:15 crc kubenswrapper[4629]: I1211 09:24:15.687363 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_147dc374-3f14-40f8-bf2f-ab1bd5a72d4e/memcached/0.log" Dec 11 09:24:20 crc kubenswrapper[4629]: I1211 09:24:20.486516 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:24:20 crc kubenswrapper[4629]: I1211 09:24:20.487181 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:24:20 crc kubenswrapper[4629]: I1211 09:24:20.487260 4629 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 09:24:20 crc kubenswrapper[4629]: I1211 09:24:20.488180 4629 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d1441cf6cef81d3d1e6dde6fc7c192a9564615532b2a9036b19516be40e634b7"} pod="openshift-machine-config-operator/machine-config-daemon-wbld9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 09:24:20 crc kubenswrapper[4629]: I1211 09:24:20.488251 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" containerID="cri-o://d1441cf6cef81d3d1e6dde6fc7c192a9564615532b2a9036b19516be40e634b7" gracePeriod=600 Dec 11 09:24:20 crc kubenswrapper[4629]: I1211 09:24:20.875980 4629 generic.go:334] "Generic (PLEG): container finished" podID="847b50b9-0093-404f-b0dd-63c814d8986b" containerID="d1441cf6cef81d3d1e6dde6fc7c192a9564615532b2a9036b19516be40e634b7" exitCode=0 Dec 11 09:24:20 crc kubenswrapper[4629]: I1211 09:24:20.876069 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerDied","Data":"d1441cf6cef81d3d1e6dde6fc7c192a9564615532b2a9036b19516be40e634b7"} Dec 11 09:24:20 crc kubenswrapper[4629]: I1211 09:24:20.876344 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c"} Dec 11 09:24:20 crc kubenswrapper[4629]: I1211 09:24:20.876367 4629 scope.go:117] "RemoveContainer" containerID="76786ba20d20bc31b0eb8f944f46ecc05ad246fcd290e374e8975b246d3b8e48" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.279108 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ghdvs"] Dec 11 09:24:29 crc kubenswrapper[4629]: E1211 09:24:29.280223 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2060d1a6-62e9-4a5e-b36c-090fb9292999" containerName="registry-server" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.280241 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="2060d1a6-62e9-4a5e-b36c-090fb9292999" containerName="registry-server" Dec 11 09:24:29 crc kubenswrapper[4629]: E1211 09:24:29.280285 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2060d1a6-62e9-4a5e-b36c-090fb9292999" containerName="extract-utilities" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.280296 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="2060d1a6-62e9-4a5e-b36c-090fb9292999" containerName="extract-utilities" Dec 11 09:24:29 crc kubenswrapper[4629]: E1211 09:24:29.280310 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2060d1a6-62e9-4a5e-b36c-090fb9292999" containerName="extract-content" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.280317 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="2060d1a6-62e9-4a5e-b36c-090fb9292999" containerName="extract-content" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.280546 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="2060d1a6-62e9-4a5e-b36c-090fb9292999" containerName="registry-server" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.282307 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.291651 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ghdvs"] Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.333577 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c82c24-796d-4e75-8538-e75db7dd93d5-utilities\") pod \"certified-operators-ghdvs\" (UID: \"86c82c24-796d-4e75-8538-e75db7dd93d5\") " pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.333720 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c82c24-796d-4e75-8538-e75db7dd93d5-catalog-content\") pod \"certified-operators-ghdvs\" (UID: \"86c82c24-796d-4e75-8538-e75db7dd93d5\") " pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.333835 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl572\" (UniqueName: \"kubernetes.io/projected/86c82c24-796d-4e75-8538-e75db7dd93d5-kube-api-access-nl572\") pod \"certified-operators-ghdvs\" (UID: \"86c82c24-796d-4e75-8538-e75db7dd93d5\") " pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.435357 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c82c24-796d-4e75-8538-e75db7dd93d5-utilities\") pod \"certified-operators-ghdvs\" (UID: \"86c82c24-796d-4e75-8538-e75db7dd93d5\") " pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.435447 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c82c24-796d-4e75-8538-e75db7dd93d5-catalog-content\") pod \"certified-operators-ghdvs\" (UID: \"86c82c24-796d-4e75-8538-e75db7dd93d5\") " pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.435544 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl572\" (UniqueName: \"kubernetes.io/projected/86c82c24-796d-4e75-8538-e75db7dd93d5-kube-api-access-nl572\") pod \"certified-operators-ghdvs\" (UID: \"86c82c24-796d-4e75-8538-e75db7dd93d5\") " pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.436487 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c82c24-796d-4e75-8538-e75db7dd93d5-catalog-content\") pod \"certified-operators-ghdvs\" (UID: \"86c82c24-796d-4e75-8538-e75db7dd93d5\") " pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.436656 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c82c24-796d-4e75-8538-e75db7dd93d5-utilities\") pod \"certified-operators-ghdvs\" (UID: \"86c82c24-796d-4e75-8538-e75db7dd93d5\") " pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.460715 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl572\" (UniqueName: \"kubernetes.io/projected/86c82c24-796d-4e75-8538-e75db7dd93d5-kube-api-access-nl572\") pod \"certified-operators-ghdvs\" (UID: \"86c82c24-796d-4e75-8538-e75db7dd93d5\") " pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:29 crc kubenswrapper[4629]: I1211 09:24:29.615879 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:30 crc kubenswrapper[4629]: I1211 09:24:30.198500 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ghdvs"] Dec 11 09:24:30 crc kubenswrapper[4629]: I1211 09:24:30.966173 4629 generic.go:334] "Generic (PLEG): container finished" podID="86c82c24-796d-4e75-8538-e75db7dd93d5" containerID="1d1879c23f03770e49dbd52d34de743d3fa28ec45beebd66f447535440ac0e58" exitCode=0 Dec 11 09:24:30 crc kubenswrapper[4629]: I1211 09:24:30.966299 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ghdvs" event={"ID":"86c82c24-796d-4e75-8538-e75db7dd93d5","Type":"ContainerDied","Data":"1d1879c23f03770e49dbd52d34de743d3fa28ec45beebd66f447535440ac0e58"} Dec 11 09:24:30 crc kubenswrapper[4629]: I1211 09:24:30.966905 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ghdvs" event={"ID":"86c82c24-796d-4e75-8538-e75db7dd93d5","Type":"ContainerStarted","Data":"f230a266d1e88d0916014742c098b6d55442e2f09c94fbb1e4c10b00560c1e30"} Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.450359 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qff8g"] Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.453465 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.463134 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qff8g"] Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.488656 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362675ee-0b61-4101-9ea8-1b07f020bc73-catalog-content\") pod \"redhat-operators-qff8g\" (UID: \"362675ee-0b61-4101-9ea8-1b07f020bc73\") " pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.488770 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mdmw\" (UniqueName: \"kubernetes.io/projected/362675ee-0b61-4101-9ea8-1b07f020bc73-kube-api-access-4mdmw\") pod \"redhat-operators-qff8g\" (UID: \"362675ee-0b61-4101-9ea8-1b07f020bc73\") " pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.488860 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362675ee-0b61-4101-9ea8-1b07f020bc73-utilities\") pod \"redhat-operators-qff8g\" (UID: \"362675ee-0b61-4101-9ea8-1b07f020bc73\") " pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.590432 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362675ee-0b61-4101-9ea8-1b07f020bc73-catalog-content\") pod \"redhat-operators-qff8g\" (UID: \"362675ee-0b61-4101-9ea8-1b07f020bc73\") " pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.590970 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mdmw\" (UniqueName: \"kubernetes.io/projected/362675ee-0b61-4101-9ea8-1b07f020bc73-kube-api-access-4mdmw\") pod \"redhat-operators-qff8g\" (UID: \"362675ee-0b61-4101-9ea8-1b07f020bc73\") " pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.591096 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362675ee-0b61-4101-9ea8-1b07f020bc73-utilities\") pod \"redhat-operators-qff8g\" (UID: \"362675ee-0b61-4101-9ea8-1b07f020bc73\") " pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.591323 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362675ee-0b61-4101-9ea8-1b07f020bc73-catalog-content\") pod \"redhat-operators-qff8g\" (UID: \"362675ee-0b61-4101-9ea8-1b07f020bc73\") " pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.591648 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362675ee-0b61-4101-9ea8-1b07f020bc73-utilities\") pod \"redhat-operators-qff8g\" (UID: \"362675ee-0b61-4101-9ea8-1b07f020bc73\") " pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.615204 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mdmw\" (UniqueName: \"kubernetes.io/projected/362675ee-0b61-4101-9ea8-1b07f020bc73-kube-api-access-4mdmw\") pod \"redhat-operators-qff8g\" (UID: \"362675ee-0b61-4101-9ea8-1b07f020bc73\") " pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.657749 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pg68h"] Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.660312 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.676093 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pg68h"] Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.692337 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3041cd3-67f1-4459-91a2-c548576aaf76-utilities\") pod \"community-operators-pg68h\" (UID: \"c3041cd3-67f1-4459-91a2-c548576aaf76\") " pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.692384 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3041cd3-67f1-4459-91a2-c548576aaf76-catalog-content\") pod \"community-operators-pg68h\" (UID: \"c3041cd3-67f1-4459-91a2-c548576aaf76\") " pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.692459 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74hsr\" (UniqueName: \"kubernetes.io/projected/c3041cd3-67f1-4459-91a2-c548576aaf76-kube-api-access-74hsr\") pod \"community-operators-pg68h\" (UID: \"c3041cd3-67f1-4459-91a2-c548576aaf76\") " pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.789156 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.794575 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74hsr\" (UniqueName: \"kubernetes.io/projected/c3041cd3-67f1-4459-91a2-c548576aaf76-kube-api-access-74hsr\") pod \"community-operators-pg68h\" (UID: \"c3041cd3-67f1-4459-91a2-c548576aaf76\") " pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.794805 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3041cd3-67f1-4459-91a2-c548576aaf76-utilities\") pod \"community-operators-pg68h\" (UID: \"c3041cd3-67f1-4459-91a2-c548576aaf76\") " pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.795488 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3041cd3-67f1-4459-91a2-c548576aaf76-utilities\") pod \"community-operators-pg68h\" (UID: \"c3041cd3-67f1-4459-91a2-c548576aaf76\") " pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.795543 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3041cd3-67f1-4459-91a2-c548576aaf76-catalog-content\") pod \"community-operators-pg68h\" (UID: \"c3041cd3-67f1-4459-91a2-c548576aaf76\") " pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.795639 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3041cd3-67f1-4459-91a2-c548576aaf76-catalog-content\") pod \"community-operators-pg68h\" (UID: \"c3041cd3-67f1-4459-91a2-c548576aaf76\") " pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.813698 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74hsr\" (UniqueName: \"kubernetes.io/projected/c3041cd3-67f1-4459-91a2-c548576aaf76-kube-api-access-74hsr\") pod \"community-operators-pg68h\" (UID: \"c3041cd3-67f1-4459-91a2-c548576aaf76\") " pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:31 crc kubenswrapper[4629]: I1211 09:24:31.987825 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:32 crc kubenswrapper[4629]: I1211 09:24:32.171348 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qff8g"] Dec 11 09:24:32 crc kubenswrapper[4629]: W1211 09:24:32.193013 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod362675ee_0b61_4101_9ea8_1b07f020bc73.slice/crio-6e8d803c5a8a1d27e1e60c3050a6d97f2c2480b7b1d298a7f3c90634bbd5d5b0 WatchSource:0}: Error finding container 6e8d803c5a8a1d27e1e60c3050a6d97f2c2480b7b1d298a7f3c90634bbd5d5b0: Status 404 returned error can't find the container with id 6e8d803c5a8a1d27e1e60c3050a6d97f2c2480b7b1d298a7f3c90634bbd5d5b0 Dec 11 09:24:32 crc kubenswrapper[4629]: W1211 09:24:32.801055 4629 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3041cd3_67f1_4459_91a2_c548576aaf76.slice/crio-8216da742de987eba3ca694de7b6023d05057db58104cc42ed63e87446f99162 WatchSource:0}: Error finding container 8216da742de987eba3ca694de7b6023d05057db58104cc42ed63e87446f99162: Status 404 returned error can't find the container with id 8216da742de987eba3ca694de7b6023d05057db58104cc42ed63e87446f99162 Dec 11 09:24:32 crc kubenswrapper[4629]: I1211 09:24:32.832375 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pg68h"] Dec 11 09:24:33 crc kubenswrapper[4629]: I1211 09:24:33.009440 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg68h" event={"ID":"c3041cd3-67f1-4459-91a2-c548576aaf76","Type":"ContainerStarted","Data":"8216da742de987eba3ca694de7b6023d05057db58104cc42ed63e87446f99162"} Dec 11 09:24:33 crc kubenswrapper[4629]: I1211 09:24:33.014308 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ghdvs" event={"ID":"86c82c24-796d-4e75-8538-e75db7dd93d5","Type":"ContainerStarted","Data":"cdb968d3c3627d19d81a7db140196f50d87c36332903654c4fc14b3192992e35"} Dec 11 09:24:33 crc kubenswrapper[4629]: I1211 09:24:33.019456 4629 generic.go:334] "Generic (PLEG): container finished" podID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerID="4e7c5e4838ab0e705e4fc37b9b885a0a8bbf62015a2523e2fc229649937a8eec" exitCode=0 Dec 11 09:24:33 crc kubenswrapper[4629]: I1211 09:24:33.019510 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qff8g" event={"ID":"362675ee-0b61-4101-9ea8-1b07f020bc73","Type":"ContainerDied","Data":"4e7c5e4838ab0e705e4fc37b9b885a0a8bbf62015a2523e2fc229649937a8eec"} Dec 11 09:24:33 crc kubenswrapper[4629]: I1211 09:24:33.019545 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qff8g" event={"ID":"362675ee-0b61-4101-9ea8-1b07f020bc73","Type":"ContainerStarted","Data":"6e8d803c5a8a1d27e1e60c3050a6d97f2c2480b7b1d298a7f3c90634bbd5d5b0"} Dec 11 09:24:34 crc kubenswrapper[4629]: I1211 09:24:34.035889 4629 generic.go:334] "Generic (PLEG): container finished" podID="c3041cd3-67f1-4459-91a2-c548576aaf76" containerID="fc065bb6c85f2ecb4f591d32170016cc1598ace86f844c50912b07cc45bc9609" exitCode=0 Dec 11 09:24:34 crc kubenswrapper[4629]: I1211 09:24:34.036123 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg68h" event={"ID":"c3041cd3-67f1-4459-91a2-c548576aaf76","Type":"ContainerDied","Data":"fc065bb6c85f2ecb4f591d32170016cc1598ace86f844c50912b07cc45bc9609"} Dec 11 09:24:34 crc kubenswrapper[4629]: I1211 09:24:34.040133 4629 generic.go:334] "Generic (PLEG): container finished" podID="86c82c24-796d-4e75-8538-e75db7dd93d5" containerID="cdb968d3c3627d19d81a7db140196f50d87c36332903654c4fc14b3192992e35" exitCode=0 Dec 11 09:24:34 crc kubenswrapper[4629]: I1211 09:24:34.040168 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ghdvs" event={"ID":"86c82c24-796d-4e75-8538-e75db7dd93d5","Type":"ContainerDied","Data":"cdb968d3c3627d19d81a7db140196f50d87c36332903654c4fc14b3192992e35"} Dec 11 09:24:35 crc kubenswrapper[4629]: I1211 09:24:35.054103 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qff8g" event={"ID":"362675ee-0b61-4101-9ea8-1b07f020bc73","Type":"ContainerStarted","Data":"fdc2afce703be924f963c93c0a1d41f76d63080b72fdd95f86847ca59d3e0f28"} Dec 11 09:24:35 crc kubenswrapper[4629]: I1211 09:24:35.058348 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg68h" event={"ID":"c3041cd3-67f1-4459-91a2-c548576aaf76","Type":"ContainerStarted","Data":"072acea5f332625a159aca198441980ff25f5eb7070a86bfcd0fd32112864d7d"} Dec 11 09:24:36 crc kubenswrapper[4629]: I1211 09:24:36.087195 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ghdvs" event={"ID":"86c82c24-796d-4e75-8538-e75db7dd93d5","Type":"ContainerStarted","Data":"e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0"} Dec 11 09:24:36 crc kubenswrapper[4629]: I1211 09:24:36.122474 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ghdvs" podStartSLOduration=2.639345554 podStartE2EDuration="7.122443335s" podCreationTimestamp="2025-12-11 09:24:29 +0000 UTC" firstStartedPulling="2025-12-11 09:24:30.970399374 +0000 UTC m=+2859.035814982" lastFinishedPulling="2025-12-11 09:24:35.453497155 +0000 UTC m=+2863.518912763" observedRunningTime="2025-12-11 09:24:36.115461895 +0000 UTC m=+2864.180877503" watchObservedRunningTime="2025-12-11 09:24:36.122443335 +0000 UTC m=+2864.187858943" Dec 11 09:24:39 crc kubenswrapper[4629]: I1211 09:24:39.117823 4629 generic.go:334] "Generic (PLEG): container finished" podID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerID="fdc2afce703be924f963c93c0a1d41f76d63080b72fdd95f86847ca59d3e0f28" exitCode=0 Dec 11 09:24:39 crc kubenswrapper[4629]: I1211 09:24:39.117895 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qff8g" event={"ID":"362675ee-0b61-4101-9ea8-1b07f020bc73","Type":"ContainerDied","Data":"fdc2afce703be924f963c93c0a1d41f76d63080b72fdd95f86847ca59d3e0f28"} Dec 11 09:24:39 crc kubenswrapper[4629]: I1211 09:24:39.127958 4629 generic.go:334] "Generic (PLEG): container finished" podID="c3041cd3-67f1-4459-91a2-c548576aaf76" containerID="072acea5f332625a159aca198441980ff25f5eb7070a86bfcd0fd32112864d7d" exitCode=0 Dec 11 09:24:39 crc kubenswrapper[4629]: I1211 09:24:39.128002 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg68h" event={"ID":"c3041cd3-67f1-4459-91a2-c548576aaf76","Type":"ContainerDied","Data":"072acea5f332625a159aca198441980ff25f5eb7070a86bfcd0fd32112864d7d"} Dec 11 09:24:39 crc kubenswrapper[4629]: I1211 09:24:39.616569 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:39 crc kubenswrapper[4629]: I1211 09:24:39.616964 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:39 crc kubenswrapper[4629]: I1211 09:24:39.670068 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:40 crc kubenswrapper[4629]: I1211 09:24:40.189414 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:40 crc kubenswrapper[4629]: I1211 09:24:40.644657 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ghdvs"] Dec 11 09:24:41 crc kubenswrapper[4629]: I1211 09:24:41.162414 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qff8g" event={"ID":"362675ee-0b61-4101-9ea8-1b07f020bc73","Type":"ContainerStarted","Data":"d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0"} Dec 11 09:24:41 crc kubenswrapper[4629]: I1211 09:24:41.167796 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg68h" event={"ID":"c3041cd3-67f1-4459-91a2-c548576aaf76","Type":"ContainerStarted","Data":"e5855caf15e7a4089d20234a8764b0d6f5c24288f84571e37b9a3560ade421c5"} Dec 11 09:24:41 crc kubenswrapper[4629]: I1211 09:24:41.192395 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qff8g" podStartSLOduration=2.904972515 podStartE2EDuration="10.192368817s" podCreationTimestamp="2025-12-11 09:24:31 +0000 UTC" firstStartedPulling="2025-12-11 09:24:33.024781054 +0000 UTC m=+2861.090196652" lastFinishedPulling="2025-12-11 09:24:40.312177336 +0000 UTC m=+2868.377592954" observedRunningTime="2025-12-11 09:24:41.185613465 +0000 UTC m=+2869.251029073" watchObservedRunningTime="2025-12-11 09:24:41.192368817 +0000 UTC m=+2869.257784425" Dec 11 09:24:41 crc kubenswrapper[4629]: I1211 09:24:41.789434 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:41 crc kubenswrapper[4629]: I1211 09:24:41.789498 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:24:41 crc kubenswrapper[4629]: I1211 09:24:41.989236 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:41 crc kubenswrapper[4629]: I1211 09:24:41.989608 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.173429 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ghdvs" podUID="86c82c24-796d-4e75-8538-e75db7dd93d5" containerName="registry-server" containerID="cri-o://e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0" gracePeriod=2 Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.670528 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.698801 4629 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pg68h" podStartSLOduration=5.391541712 podStartE2EDuration="11.698779925s" podCreationTimestamp="2025-12-11 09:24:31 +0000 UTC" firstStartedPulling="2025-12-11 09:24:34.038696623 +0000 UTC m=+2862.104112231" lastFinishedPulling="2025-12-11 09:24:40.345934836 +0000 UTC m=+2868.411350444" observedRunningTime="2025-12-11 09:24:41.216701991 +0000 UTC m=+2869.282117599" watchObservedRunningTime="2025-12-11 09:24:42.698779925 +0000 UTC m=+2870.764195523" Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.708150 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c82c24-796d-4e75-8538-e75db7dd93d5-utilities\") pod \"86c82c24-796d-4e75-8538-e75db7dd93d5\" (UID: \"86c82c24-796d-4e75-8538-e75db7dd93d5\") " Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.708272 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c82c24-796d-4e75-8538-e75db7dd93d5-catalog-content\") pod \"86c82c24-796d-4e75-8538-e75db7dd93d5\" (UID: \"86c82c24-796d-4e75-8538-e75db7dd93d5\") " Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.708406 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nl572\" (UniqueName: \"kubernetes.io/projected/86c82c24-796d-4e75-8538-e75db7dd93d5-kube-api-access-nl572\") pod \"86c82c24-796d-4e75-8538-e75db7dd93d5\" (UID: \"86c82c24-796d-4e75-8538-e75db7dd93d5\") " Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.711174 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86c82c24-796d-4e75-8538-e75db7dd93d5-utilities" (OuterVolumeSpecName: "utilities") pod "86c82c24-796d-4e75-8538-e75db7dd93d5" (UID: "86c82c24-796d-4e75-8538-e75db7dd93d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.732283 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86c82c24-796d-4e75-8538-e75db7dd93d5-kube-api-access-nl572" (OuterVolumeSpecName: "kube-api-access-nl572") pod "86c82c24-796d-4e75-8538-e75db7dd93d5" (UID: "86c82c24-796d-4e75-8538-e75db7dd93d5"). InnerVolumeSpecName "kube-api-access-nl572". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.785196 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86c82c24-796d-4e75-8538-e75db7dd93d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86c82c24-796d-4e75-8538-e75db7dd93d5" (UID: "86c82c24-796d-4e75-8538-e75db7dd93d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.811058 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c82c24-796d-4e75-8538-e75db7dd93d5-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.811101 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c82c24-796d-4e75-8538-e75db7dd93d5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.811117 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nl572\" (UniqueName: \"kubernetes.io/projected/86c82c24-796d-4e75-8538-e75db7dd93d5-kube-api-access-nl572\") on node \"crc\" DevicePath \"\"" Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.853270 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qff8g" podUID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerName="registry-server" probeResult="failure" output=< Dec 11 09:24:42 crc kubenswrapper[4629]: timeout: failed to connect service ":50051" within 1s Dec 11 09:24:42 crc kubenswrapper[4629]: > Dec 11 09:24:42 crc kubenswrapper[4629]: I1211 09:24:42.898418 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v_6efefc09-794d-479a-b21a-b1c43506bcae/util/0.log" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.058142 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-pg68h" podUID="c3041cd3-67f1-4459-91a2-c548576aaf76" containerName="registry-server" probeResult="failure" output=< Dec 11 09:24:43 crc kubenswrapper[4629]: timeout: failed to connect service ":50051" within 1s Dec 11 09:24:43 crc kubenswrapper[4629]: > Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.087818 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v_6efefc09-794d-479a-b21a-b1c43506bcae/pull/0.log" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.152362 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v_6efefc09-794d-479a-b21a-b1c43506bcae/pull/0.log" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.186130 4629 generic.go:334] "Generic (PLEG): container finished" podID="86c82c24-796d-4e75-8538-e75db7dd93d5" containerID="e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0" exitCode=0 Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.187235 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ghdvs" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.187420 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ghdvs" event={"ID":"86c82c24-796d-4e75-8538-e75db7dd93d5","Type":"ContainerDied","Data":"e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0"} Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.187468 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ghdvs" event={"ID":"86c82c24-796d-4e75-8538-e75db7dd93d5","Type":"ContainerDied","Data":"f230a266d1e88d0916014742c098b6d55442e2f09c94fbb1e4c10b00560c1e30"} Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.187492 4629 scope.go:117] "RemoveContainer" containerID="e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.204533 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v_6efefc09-794d-479a-b21a-b1c43506bcae/util/0.log" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.214659 4629 scope.go:117] "RemoveContainer" containerID="cdb968d3c3627d19d81a7db140196f50d87c36332903654c4fc14b3192992e35" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.245271 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ghdvs"] Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.268732 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ghdvs"] Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.279657 4629 scope.go:117] "RemoveContainer" containerID="1d1879c23f03770e49dbd52d34de743d3fa28ec45beebd66f447535440ac0e58" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.299545 4629 scope.go:117] "RemoveContainer" containerID="e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0" Dec 11 09:24:43 crc kubenswrapper[4629]: E1211 09:24:43.303154 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0\": container with ID starting with e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0 not found: ID does not exist" containerID="e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.303251 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0"} err="failed to get container status \"e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0\": rpc error: code = NotFound desc = could not find container \"e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0\": container with ID starting with e806c5e9eb8c106624b4112852f76f9240f9118e97b7907835398ada61c0e6e0 not found: ID does not exist" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.303287 4629 scope.go:117] "RemoveContainer" containerID="cdb968d3c3627d19d81a7db140196f50d87c36332903654c4fc14b3192992e35" Dec 11 09:24:43 crc kubenswrapper[4629]: E1211 09:24:43.304238 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdb968d3c3627d19d81a7db140196f50d87c36332903654c4fc14b3192992e35\": container with ID starting with cdb968d3c3627d19d81a7db140196f50d87c36332903654c4fc14b3192992e35 not found: ID does not exist" containerID="cdb968d3c3627d19d81a7db140196f50d87c36332903654c4fc14b3192992e35" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.304312 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdb968d3c3627d19d81a7db140196f50d87c36332903654c4fc14b3192992e35"} err="failed to get container status \"cdb968d3c3627d19d81a7db140196f50d87c36332903654c4fc14b3192992e35\": rpc error: code = NotFound desc = could not find container \"cdb968d3c3627d19d81a7db140196f50d87c36332903654c4fc14b3192992e35\": container with ID starting with cdb968d3c3627d19d81a7db140196f50d87c36332903654c4fc14b3192992e35 not found: ID does not exist" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.304347 4629 scope.go:117] "RemoveContainer" containerID="1d1879c23f03770e49dbd52d34de743d3fa28ec45beebd66f447535440ac0e58" Dec 11 09:24:43 crc kubenswrapper[4629]: E1211 09:24:43.304806 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d1879c23f03770e49dbd52d34de743d3fa28ec45beebd66f447535440ac0e58\": container with ID starting with 1d1879c23f03770e49dbd52d34de743d3fa28ec45beebd66f447535440ac0e58 not found: ID does not exist" containerID="1d1879c23f03770e49dbd52d34de743d3fa28ec45beebd66f447535440ac0e58" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.304871 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d1879c23f03770e49dbd52d34de743d3fa28ec45beebd66f447535440ac0e58"} err="failed to get container status \"1d1879c23f03770e49dbd52d34de743d3fa28ec45beebd66f447535440ac0e58\": rpc error: code = NotFound desc = could not find container \"1d1879c23f03770e49dbd52d34de743d3fa28ec45beebd66f447535440ac0e58\": container with ID starting with 1d1879c23f03770e49dbd52d34de743d3fa28ec45beebd66f447535440ac0e58 not found: ID does not exist" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.491960 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v_6efefc09-794d-479a-b21a-b1c43506bcae/util/0.log" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.533883 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v_6efefc09-794d-479a-b21a-b1c43506bcae/pull/0.log" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.618934 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3dedefb9ef33010c8c547f48a376dfbf037b83356e9bd17162bcfc27c6qlc9v_6efefc09-794d-479a-b21a-b1c43506bcae/extract/0.log" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.735427 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-g5zpd_813d511c-6f11-40c8-a9c3-281e4909c88d/kube-rbac-proxy/0.log" Dec 11 09:24:43 crc kubenswrapper[4629]: I1211 09:24:43.908033 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-g5zpd_813d511c-6f11-40c8-a9c3-281e4909c88d/manager/0.log" Dec 11 09:24:44 crc kubenswrapper[4629]: I1211 09:24:44.006823 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-p8lh9_39798d73-c400-4797-b5ac-e402d17a1df1/kube-rbac-proxy/0.log" Dec 11 09:24:44 crc kubenswrapper[4629]: I1211 09:24:44.208799 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86c82c24-796d-4e75-8538-e75db7dd93d5" path="/var/lib/kubelet/pods/86c82c24-796d-4e75-8538-e75db7dd93d5/volumes" Dec 11 09:24:44 crc kubenswrapper[4629]: I1211 09:24:44.226438 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-mhqnx_c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1/manager/0.log" Dec 11 09:24:44 crc kubenswrapper[4629]: I1211 09:24:44.307752 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-mhqnx_c7bac7ad-d9f6-4d34-bcab-3bbc2d5e95f1/kube-rbac-proxy/0.log" Dec 11 09:24:44 crc kubenswrapper[4629]: I1211 09:24:44.308746 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-p8lh9_39798d73-c400-4797-b5ac-e402d17a1df1/manager/0.log" Dec 11 09:24:44 crc kubenswrapper[4629]: I1211 09:24:44.532140 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-xdp6j_64242f6c-6d25-4868-acd8-de067ecdeb7a/kube-rbac-proxy/0.log" Dec 11 09:24:44 crc kubenswrapper[4629]: I1211 09:24:44.574669 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-xdp6j_64242f6c-6d25-4868-acd8-de067ecdeb7a/manager/0.log" Dec 11 09:24:44 crc kubenswrapper[4629]: I1211 09:24:44.724586 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-8mtlt_8a37b92a-08fc-47a9-8d55-97182286daea/kube-rbac-proxy/0.log" Dec 11 09:24:44 crc kubenswrapper[4629]: I1211 09:24:44.784261 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-8mtlt_8a37b92a-08fc-47a9-8d55-97182286daea/manager/0.log" Dec 11 09:24:44 crc kubenswrapper[4629]: I1211 09:24:44.879087 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kq7vj_7833d1b6-8dbd-4afa-9042-755b5a9c383b/kube-rbac-proxy/0.log" Dec 11 09:24:45 crc kubenswrapper[4629]: I1211 09:24:45.020042 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kq7vj_7833d1b6-8dbd-4afa-9042-755b5a9c383b/manager/0.log" Dec 11 09:24:45 crc kubenswrapper[4629]: I1211 09:24:45.079953 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-l59gh_219ddd48-017d-4da4-b576-0b8f24bc40be/kube-rbac-proxy/0.log" Dec 11 09:24:45 crc kubenswrapper[4629]: I1211 09:24:45.364177 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-l59gh_219ddd48-017d-4da4-b576-0b8f24bc40be/manager/0.log" Dec 11 09:24:45 crc kubenswrapper[4629]: I1211 09:24:45.418603 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-qxsl5_bf19951f-862f-4b74-92ab-63cdede5be6d/manager/0.log" Dec 11 09:24:45 crc kubenswrapper[4629]: I1211 09:24:45.434986 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-qxsl5_bf19951f-862f-4b74-92ab-63cdede5be6d/kube-rbac-proxy/0.log" Dec 11 09:24:45 crc kubenswrapper[4629]: I1211 09:24:45.768189 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-kqcvw_94ec218d-3b08-463d-8225-f12cbcf0eb8d/manager/0.log" Dec 11 09:24:45 crc kubenswrapper[4629]: I1211 09:24:45.774465 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-kqcvw_94ec218d-3b08-463d-8225-f12cbcf0eb8d/kube-rbac-proxy/0.log" Dec 11 09:24:46 crc kubenswrapper[4629]: I1211 09:24:46.021459 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-rmlvk_bf9c412a-98bd-4152-b65a-637822c14b67/kube-rbac-proxy/0.log" Dec 11 09:24:46 crc kubenswrapper[4629]: I1211 09:24:46.031388 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-9m9lh_0f40457d-36a2-4183-8434-9cb4688489bb/manager/0.log" Dec 11 09:24:46 crc kubenswrapper[4629]: I1211 09:24:46.049381 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-9m9lh_0f40457d-36a2-4183-8434-9cb4688489bb/kube-rbac-proxy/0.log" Dec 11 09:24:46 crc kubenswrapper[4629]: I1211 09:24:46.374080 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-rmlvk_bf9c412a-98bd-4152-b65a-637822c14b67/manager/0.log" Dec 11 09:24:46 crc kubenswrapper[4629]: I1211 09:24:46.396481 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-mjhzk_8c62910d-1b99-40df-be12-c3eb86c645aa/kube-rbac-proxy/0.log" Dec 11 09:24:46 crc kubenswrapper[4629]: I1211 09:24:46.436316 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-mjhzk_8c62910d-1b99-40df-be12-c3eb86c645aa/manager/0.log" Dec 11 09:24:46 crc kubenswrapper[4629]: I1211 09:24:46.654167 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-p9wjh_99ac9552-d81c-47b9-a77c-c18fae357872/kube-rbac-proxy/0.log" Dec 11 09:24:46 crc kubenswrapper[4629]: I1211 09:24:46.748531 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-p9wjh_99ac9552-d81c-47b9-a77c-c18fae357872/manager/0.log" Dec 11 09:24:46 crc kubenswrapper[4629]: I1211 09:24:46.886479 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-tg2w6_3daa4e0a-c804-4d70-a79c-5e4d5dd3a602/manager/0.log" Dec 11 09:24:46 crc kubenswrapper[4629]: I1211 09:24:46.897463 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-tg2w6_3daa4e0a-c804-4d70-a79c-5e4d5dd3a602/kube-rbac-proxy/0.log" Dec 11 09:24:47 crc kubenswrapper[4629]: I1211 09:24:47.096628 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fr7wqh_49030eaa-0443-4db5-bfd2-3c4e17a3780e/kube-rbac-proxy/0.log" Dec 11 09:24:47 crc kubenswrapper[4629]: I1211 09:24:47.158637 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fr7wqh_49030eaa-0443-4db5-bfd2-3c4e17a3780e/manager/0.log" Dec 11 09:24:47 crc kubenswrapper[4629]: I1211 09:24:47.530027 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-kgkwp_1058bc3d-a814-4f48-b25f-d6750f55d459/registry-server/0.log" Dec 11 09:24:47 crc kubenswrapper[4629]: I1211 09:24:47.630088 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7b8547f8bd-jz7pk_9dc7763f-4b47-4337-b320-fc46f74c2ed3/operator/0.log" Dec 11 09:24:47 crc kubenswrapper[4629]: I1211 09:24:47.856456 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-f5ftx_c808480f-8564-4628-b9b9-7ee9700cbb91/kube-rbac-proxy/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.027429 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-f5ftx_c808480f-8564-4628-b9b9-7ee9700cbb91/manager/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.083254 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-8q2kk_8e2cace8-e27c-4d40-b20e-dd16cb10a425/kube-rbac-proxy/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.157995 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-8q2kk_8e2cace8-e27c-4d40-b20e-dd16cb10a425/manager/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.209000 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-69b5bb5b-x6z9b_914206a5-adb8-4482-90f6-e55a294259c3/manager/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.393161 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-kvzwf_5af0ef89-5415-482d-bbe9-d97cb6da58d6/operator/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.426801 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-gtqck_8ebecdcb-ee18-4285-8b09-65c8859dc77e/kube-rbac-proxy/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.456667 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-gtqck_8ebecdcb-ee18-4285-8b09-65c8859dc77e/manager/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.569159 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-cptvh_1150e5d5-ae4c-4800-a10d-9ba271074338/kube-rbac-proxy/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.658642 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-cptvh_1150e5d5-ae4c-4800-a10d-9ba271074338/manager/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.695606 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-f8m45_2382ec97-3dac-4285-bb95-fa1864b1bd8d/kube-rbac-proxy/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.834283 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-f8m45_2382ec97-3dac-4285-bb95-fa1864b1bd8d/manager/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.926310 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-8qz72_a6ef2a86-805b-44ed-95bd-1c53e457b9d9/kube-rbac-proxy/0.log" Dec 11 09:24:48 crc kubenswrapper[4629]: I1211 09:24:48.934150 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-8qz72_a6ef2a86-805b-44ed-95bd-1c53e457b9d9/manager/0.log" Dec 11 09:24:52 crc kubenswrapper[4629]: I1211 09:24:52.041832 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:52 crc kubenswrapper[4629]: I1211 09:24:52.098650 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:52 crc kubenswrapper[4629]: I1211 09:24:52.836553 4629 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qff8g" podUID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerName="registry-server" probeResult="failure" output=< Dec 11 09:24:52 crc kubenswrapper[4629]: timeout: failed to connect service ":50051" within 1s Dec 11 09:24:52 crc kubenswrapper[4629]: > Dec 11 09:24:53 crc kubenswrapper[4629]: I1211 09:24:53.278758 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pg68h"] Dec 11 09:24:53 crc kubenswrapper[4629]: I1211 09:24:53.280106 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pg68h" podUID="c3041cd3-67f1-4459-91a2-c548576aaf76" containerName="registry-server" containerID="cri-o://e5855caf15e7a4089d20234a8764b0d6f5c24288f84571e37b9a3560ade421c5" gracePeriod=2 Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.291862 4629 generic.go:334] "Generic (PLEG): container finished" podID="c3041cd3-67f1-4459-91a2-c548576aaf76" containerID="e5855caf15e7a4089d20234a8764b0d6f5c24288f84571e37b9a3560ade421c5" exitCode=0 Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.291916 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg68h" event={"ID":"c3041cd3-67f1-4459-91a2-c548576aaf76","Type":"ContainerDied","Data":"e5855caf15e7a4089d20234a8764b0d6f5c24288f84571e37b9a3560ade421c5"} Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.292441 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg68h" event={"ID":"c3041cd3-67f1-4459-91a2-c548576aaf76","Type":"ContainerDied","Data":"8216da742de987eba3ca694de7b6023d05057db58104cc42ed63e87446f99162"} Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.292458 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8216da742de987eba3ca694de7b6023d05057db58104cc42ed63e87446f99162" Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.334595 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.342904 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3041cd3-67f1-4459-91a2-c548576aaf76-catalog-content\") pod \"c3041cd3-67f1-4459-91a2-c548576aaf76\" (UID: \"c3041cd3-67f1-4459-91a2-c548576aaf76\") " Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.343009 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3041cd3-67f1-4459-91a2-c548576aaf76-utilities\") pod \"c3041cd3-67f1-4459-91a2-c548576aaf76\" (UID: \"c3041cd3-67f1-4459-91a2-c548576aaf76\") " Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.343080 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74hsr\" (UniqueName: \"kubernetes.io/projected/c3041cd3-67f1-4459-91a2-c548576aaf76-kube-api-access-74hsr\") pod \"c3041cd3-67f1-4459-91a2-c548576aaf76\" (UID: \"c3041cd3-67f1-4459-91a2-c548576aaf76\") " Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.343902 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3041cd3-67f1-4459-91a2-c548576aaf76-utilities" (OuterVolumeSpecName: "utilities") pod "c3041cd3-67f1-4459-91a2-c548576aaf76" (UID: "c3041cd3-67f1-4459-91a2-c548576aaf76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.349046 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3041cd3-67f1-4459-91a2-c548576aaf76-kube-api-access-74hsr" (OuterVolumeSpecName: "kube-api-access-74hsr") pod "c3041cd3-67f1-4459-91a2-c548576aaf76" (UID: "c3041cd3-67f1-4459-91a2-c548576aaf76"). InnerVolumeSpecName "kube-api-access-74hsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.398070 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3041cd3-67f1-4459-91a2-c548576aaf76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3041cd3-67f1-4459-91a2-c548576aaf76" (UID: "c3041cd3-67f1-4459-91a2-c548576aaf76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.445013 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3041cd3-67f1-4459-91a2-c548576aaf76-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.445048 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74hsr\" (UniqueName: \"kubernetes.io/projected/c3041cd3-67f1-4459-91a2-c548576aaf76-kube-api-access-74hsr\") on node \"crc\" DevicePath \"\"" Dec 11 09:24:54 crc kubenswrapper[4629]: I1211 09:24:54.445060 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3041cd3-67f1-4459-91a2-c548576aaf76-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:24:55 crc kubenswrapper[4629]: I1211 09:24:55.299808 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pg68h" Dec 11 09:24:55 crc kubenswrapper[4629]: I1211 09:24:55.394355 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pg68h"] Dec 11 09:24:55 crc kubenswrapper[4629]: I1211 09:24:55.405458 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pg68h"] Dec 11 09:24:56 crc kubenswrapper[4629]: I1211 09:24:56.208896 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3041cd3-67f1-4459-91a2-c548576aaf76" path="/var/lib/kubelet/pods/c3041cd3-67f1-4459-91a2-c548576aaf76/volumes" Dec 11 09:25:01 crc kubenswrapper[4629]: I1211 09:25:01.850825 4629 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:25:01 crc kubenswrapper[4629]: I1211 09:25:01.909235 4629 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:25:02 crc kubenswrapper[4629]: I1211 09:25:02.654272 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qff8g"] Dec 11 09:25:03 crc kubenswrapper[4629]: I1211 09:25:03.384306 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qff8g" podUID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerName="registry-server" containerID="cri-o://d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0" gracePeriod=2 Dec 11 09:25:03 crc kubenswrapper[4629]: E1211 09:25:03.646198 4629 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod362675ee_0b61_4101_9ea8_1b07f020bc73.slice/crio-conmon-d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod362675ee_0b61_4101_9ea8_1b07f020bc73.slice/crio-d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0.scope\": RecentStats: unable to find data in memory cache]" Dec 11 09:25:03 crc kubenswrapper[4629]: I1211 09:25:03.908709 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.022670 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mdmw\" (UniqueName: \"kubernetes.io/projected/362675ee-0b61-4101-9ea8-1b07f020bc73-kube-api-access-4mdmw\") pod \"362675ee-0b61-4101-9ea8-1b07f020bc73\" (UID: \"362675ee-0b61-4101-9ea8-1b07f020bc73\") " Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.022947 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362675ee-0b61-4101-9ea8-1b07f020bc73-utilities\") pod \"362675ee-0b61-4101-9ea8-1b07f020bc73\" (UID: \"362675ee-0b61-4101-9ea8-1b07f020bc73\") " Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.023106 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362675ee-0b61-4101-9ea8-1b07f020bc73-catalog-content\") pod \"362675ee-0b61-4101-9ea8-1b07f020bc73\" (UID: \"362675ee-0b61-4101-9ea8-1b07f020bc73\") " Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.023902 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/362675ee-0b61-4101-9ea8-1b07f020bc73-utilities" (OuterVolumeSpecName: "utilities") pod "362675ee-0b61-4101-9ea8-1b07f020bc73" (UID: "362675ee-0b61-4101-9ea8-1b07f020bc73"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.036287 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/362675ee-0b61-4101-9ea8-1b07f020bc73-kube-api-access-4mdmw" (OuterVolumeSpecName: "kube-api-access-4mdmw") pod "362675ee-0b61-4101-9ea8-1b07f020bc73" (UID: "362675ee-0b61-4101-9ea8-1b07f020bc73"). InnerVolumeSpecName "kube-api-access-4mdmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.036671 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mdmw\" (UniqueName: \"kubernetes.io/projected/362675ee-0b61-4101-9ea8-1b07f020bc73-kube-api-access-4mdmw\") on node \"crc\" DevicePath \"\"" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.036703 4629 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362675ee-0b61-4101-9ea8-1b07f020bc73-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.152547 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/362675ee-0b61-4101-9ea8-1b07f020bc73-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "362675ee-0b61-4101-9ea8-1b07f020bc73" (UID: "362675ee-0b61-4101-9ea8-1b07f020bc73"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.241513 4629 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362675ee-0b61-4101-9ea8-1b07f020bc73-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.394735 4629 generic.go:334] "Generic (PLEG): container finished" podID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerID="d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0" exitCode=0 Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.394780 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qff8g" event={"ID":"362675ee-0b61-4101-9ea8-1b07f020bc73","Type":"ContainerDied","Data":"d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0"} Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.394814 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qff8g" event={"ID":"362675ee-0b61-4101-9ea8-1b07f020bc73","Type":"ContainerDied","Data":"6e8d803c5a8a1d27e1e60c3050a6d97f2c2480b7b1d298a7f3c90634bbd5d5b0"} Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.394833 4629 scope.go:117] "RemoveContainer" containerID="d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.396156 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qff8g" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.421170 4629 scope.go:117] "RemoveContainer" containerID="fdc2afce703be924f963c93c0a1d41f76d63080b72fdd95f86847ca59d3e0f28" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.445925 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qff8g"] Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.462733 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qff8g"] Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.491061 4629 scope.go:117] "RemoveContainer" containerID="4e7c5e4838ab0e705e4fc37b9b885a0a8bbf62015a2523e2fc229649937a8eec" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.534541 4629 scope.go:117] "RemoveContainer" containerID="d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0" Dec 11 09:25:04 crc kubenswrapper[4629]: E1211 09:25:04.535090 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0\": container with ID starting with d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0 not found: ID does not exist" containerID="d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.535140 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0"} err="failed to get container status \"d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0\": rpc error: code = NotFound desc = could not find container \"d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0\": container with ID starting with d8b6e7811013f200c1fa6229da2d53f502fe2b046740c54511157bfa1c6010b0 not found: ID does not exist" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.535170 4629 scope.go:117] "RemoveContainer" containerID="fdc2afce703be924f963c93c0a1d41f76d63080b72fdd95f86847ca59d3e0f28" Dec 11 09:25:04 crc kubenswrapper[4629]: E1211 09:25:04.535509 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdc2afce703be924f963c93c0a1d41f76d63080b72fdd95f86847ca59d3e0f28\": container with ID starting with fdc2afce703be924f963c93c0a1d41f76d63080b72fdd95f86847ca59d3e0f28 not found: ID does not exist" containerID="fdc2afce703be924f963c93c0a1d41f76d63080b72fdd95f86847ca59d3e0f28" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.535544 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdc2afce703be924f963c93c0a1d41f76d63080b72fdd95f86847ca59d3e0f28"} err="failed to get container status \"fdc2afce703be924f963c93c0a1d41f76d63080b72fdd95f86847ca59d3e0f28\": rpc error: code = NotFound desc = could not find container \"fdc2afce703be924f963c93c0a1d41f76d63080b72fdd95f86847ca59d3e0f28\": container with ID starting with fdc2afce703be924f963c93c0a1d41f76d63080b72fdd95f86847ca59d3e0f28 not found: ID does not exist" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.535563 4629 scope.go:117] "RemoveContainer" containerID="4e7c5e4838ab0e705e4fc37b9b885a0a8bbf62015a2523e2fc229649937a8eec" Dec 11 09:25:04 crc kubenswrapper[4629]: E1211 09:25:04.535828 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e7c5e4838ab0e705e4fc37b9b885a0a8bbf62015a2523e2fc229649937a8eec\": container with ID starting with 4e7c5e4838ab0e705e4fc37b9b885a0a8bbf62015a2523e2fc229649937a8eec not found: ID does not exist" containerID="4e7c5e4838ab0e705e4fc37b9b885a0a8bbf62015a2523e2fc229649937a8eec" Dec 11 09:25:04 crc kubenswrapper[4629]: I1211 09:25:04.535875 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e7c5e4838ab0e705e4fc37b9b885a0a8bbf62015a2523e2fc229649937a8eec"} err="failed to get container status \"4e7c5e4838ab0e705e4fc37b9b885a0a8bbf62015a2523e2fc229649937a8eec\": rpc error: code = NotFound desc = could not find container \"4e7c5e4838ab0e705e4fc37b9b885a0a8bbf62015a2523e2fc229649937a8eec\": container with ID starting with 4e7c5e4838ab0e705e4fc37b9b885a0a8bbf62015a2523e2fc229649937a8eec not found: ID does not exist" Dec 11 09:25:06 crc kubenswrapper[4629]: I1211 09:25:06.215097 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="362675ee-0b61-4101-9ea8-1b07f020bc73" path="/var/lib/kubelet/pods/362675ee-0b61-4101-9ea8-1b07f020bc73/volumes" Dec 11 09:25:12 crc kubenswrapper[4629]: I1211 09:25:12.251830 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-m9222_aedfd32e-f597-4753-8fbc-824b562b9c77/control-plane-machine-set-operator/0.log" Dec 11 09:25:12 crc kubenswrapper[4629]: I1211 09:25:12.459279 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-npk4p_ed370895-f652-499d-8382-69fb05f78f03/kube-rbac-proxy/0.log" Dec 11 09:25:12 crc kubenswrapper[4629]: I1211 09:25:12.480957 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-npk4p_ed370895-f652-499d-8382-69fb05f78f03/machine-api-operator/0.log" Dec 11 09:25:26 crc kubenswrapper[4629]: I1211 09:25:26.721635 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nplz8_0fdee0b6-4a25-4531-bca0-adae607ea9c2/cert-manager-controller/0.log" Dec 11 09:25:26 crc kubenswrapper[4629]: I1211 09:25:26.877381 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-f4zfd_5127d937-f916-4b39-bec4-79febeaeb994/cert-manager-cainjector/0.log" Dec 11 09:25:26 crc kubenswrapper[4629]: I1211 09:25:26.972937 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-8td5b_17c0e2d5-d136-4e6b-a164-c43444ba024d/cert-manager-webhook/0.log" Dec 11 09:25:41 crc kubenswrapper[4629]: I1211 09:25:41.085998 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6ff7998486-nm6nc_d89bee79-f8ff-4b46-98d0-b76671baf508/nmstate-console-plugin/0.log" Dec 11 09:25:41 crc kubenswrapper[4629]: I1211 09:25:41.422191 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-t7j8b_80988eb2-1173-47de-8699-b7e7bd1f63b7/nmstate-handler/0.log" Dec 11 09:25:41 crc kubenswrapper[4629]: I1211 09:25:41.452318 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-f64ct_39b77803-5bb8-436f-8ca6-bc4297ea81e8/nmstate-metrics/0.log" Dec 11 09:25:41 crc kubenswrapper[4629]: I1211 09:25:41.452506 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-f64ct_39b77803-5bb8-436f-8ca6-bc4297ea81e8/kube-rbac-proxy/0.log" Dec 11 09:25:41 crc kubenswrapper[4629]: I1211 09:25:41.740524 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-6769fb99d-kbnnp_e6f751dc-5c3e-4426-a89d-a38a936092e3/nmstate-operator/0.log" Dec 11 09:25:41 crc kubenswrapper[4629]: I1211 09:25:41.788953 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-f8fb84555-vtqqs_4c5aa8ca-24cf-4ac8-937d-349e4d15eae2/nmstate-webhook/0.log" Dec 11 09:25:56 crc kubenswrapper[4629]: I1211 09:25:56.300564 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-ftrtv_3fdd0751-6dee-4183-b640-85941bc266b4/kube-rbac-proxy/0.log" Dec 11 09:25:56 crc kubenswrapper[4629]: I1211 09:25:56.310827 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-ftrtv_3fdd0751-6dee-4183-b640-85941bc266b4/controller/0.log" Dec 11 09:25:56 crc kubenswrapper[4629]: I1211 09:25:56.511003 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/cp-frr-files/0.log" Dec 11 09:25:56 crc kubenswrapper[4629]: I1211 09:25:56.720539 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/cp-frr-files/0.log" Dec 11 09:25:56 crc kubenswrapper[4629]: I1211 09:25:56.790224 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/cp-reloader/0.log" Dec 11 09:25:56 crc kubenswrapper[4629]: I1211 09:25:56.807718 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/cp-reloader/0.log" Dec 11 09:25:56 crc kubenswrapper[4629]: I1211 09:25:56.819861 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/cp-metrics/0.log" Dec 11 09:25:56 crc kubenswrapper[4629]: I1211 09:25:56.965453 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/cp-frr-files/0.log" Dec 11 09:25:57 crc kubenswrapper[4629]: I1211 09:25:57.024177 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/cp-reloader/0.log" Dec 11 09:25:57 crc kubenswrapper[4629]: I1211 09:25:57.029480 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/cp-metrics/0.log" Dec 11 09:25:57 crc kubenswrapper[4629]: I1211 09:25:57.088134 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/cp-metrics/0.log" Dec 11 09:25:57 crc kubenswrapper[4629]: I1211 09:25:57.258045 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/cp-frr-files/0.log" Dec 11 09:25:57 crc kubenswrapper[4629]: I1211 09:25:57.294040 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/cp-metrics/0.log" Dec 11 09:25:57 crc kubenswrapper[4629]: I1211 09:25:57.345020 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/cp-reloader/0.log" Dec 11 09:25:57 crc kubenswrapper[4629]: I1211 09:25:57.366710 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/controller/0.log" Dec 11 09:25:57 crc kubenswrapper[4629]: I1211 09:25:57.524118 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/frr-metrics/0.log" Dec 11 09:25:57 crc kubenswrapper[4629]: I1211 09:25:57.559526 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/kube-rbac-proxy/0.log" Dec 11 09:25:57 crc kubenswrapper[4629]: I1211 09:25:57.645200 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/kube-rbac-proxy-frr/0.log" Dec 11 09:25:57 crc kubenswrapper[4629]: I1211 09:25:57.832821 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/reloader/0.log" Dec 11 09:25:58 crc kubenswrapper[4629]: I1211 09:25:58.012927 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-tf2sz_1c7c31d6-3370-4d14-861b-552a83ce767e/frr-k8s-webhook-server/0.log" Dec 11 09:25:58 crc kubenswrapper[4629]: I1211 09:25:58.368220 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6f5496ff8f-p5ldj_b51a6d95-cf61-40c5-9d03-92c5c91c16ab/manager/0.log" Dec 11 09:25:58 crc kubenswrapper[4629]: I1211 09:25:58.472656 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t6bzz_634a17de-f35e-45c3-ac4e-70b3c1768dac/frr/0.log" Dec 11 09:25:58 crc kubenswrapper[4629]: I1211 09:25:58.626090 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-649886db6b-999qk_210c22a1-a398-48f1-891f-21fecd9f53d5/webhook-server/0.log" Dec 11 09:25:58 crc kubenswrapper[4629]: I1211 09:25:58.647782 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-z8mqj_8c79723c-cb8e-4bc9-a088-2881a94dabbc/kube-rbac-proxy/0.log" Dec 11 09:25:59 crc kubenswrapper[4629]: I1211 09:25:59.008601 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-z8mqj_8c79723c-cb8e-4bc9-a088-2881a94dabbc/speaker/0.log" Dec 11 09:26:13 crc kubenswrapper[4629]: I1211 09:26:13.371479 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs_dd7dc11b-1ceb-45c3-bd58-ef07d3da9377/util/0.log" Dec 11 09:26:13 crc kubenswrapper[4629]: I1211 09:26:13.766946 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs_dd7dc11b-1ceb-45c3-bd58-ef07d3da9377/pull/0.log" Dec 11 09:26:13 crc kubenswrapper[4629]: I1211 09:26:13.863258 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs_dd7dc11b-1ceb-45c3-bd58-ef07d3da9377/util/0.log" Dec 11 09:26:13 crc kubenswrapper[4629]: I1211 09:26:13.939623 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs_dd7dc11b-1ceb-45c3-bd58-ef07d3da9377/pull/0.log" Dec 11 09:26:14 crc kubenswrapper[4629]: I1211 09:26:14.063067 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs_dd7dc11b-1ceb-45c3-bd58-ef07d3da9377/pull/0.log" Dec 11 09:26:14 crc kubenswrapper[4629]: I1211 09:26:14.091390 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs_dd7dc11b-1ceb-45c3-bd58-ef07d3da9377/util/0.log" Dec 11 09:26:14 crc kubenswrapper[4629]: I1211 09:26:14.126228 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4q2xvs_dd7dc11b-1ceb-45c3-bd58-ef07d3da9377/extract/0.log" Dec 11 09:26:14 crc kubenswrapper[4629]: I1211 09:26:14.288590 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs_a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9/util/0.log" Dec 11 09:26:14 crc kubenswrapper[4629]: I1211 09:26:14.519442 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs_a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9/pull/0.log" Dec 11 09:26:14 crc kubenswrapper[4629]: I1211 09:26:14.548432 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs_a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9/pull/0.log" Dec 11 09:26:14 crc kubenswrapper[4629]: I1211 09:26:14.552944 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs_a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9/util/0.log" Dec 11 09:26:14 crc kubenswrapper[4629]: I1211 09:26:14.764566 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs_a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9/extract/0.log" Dec 11 09:26:14 crc kubenswrapper[4629]: I1211 09:26:14.795539 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs_a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9/util/0.log" Dec 11 09:26:14 crc kubenswrapper[4629]: I1211 09:26:14.886732 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8x8rbs_a5348fbc-72ec-4ed1-8b77-529b0c9ea5b9/pull/0.log" Dec 11 09:26:14 crc kubenswrapper[4629]: I1211 09:26:14.985677 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bv9ls_d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130/extract-utilities/0.log" Dec 11 09:26:15 crc kubenswrapper[4629]: I1211 09:26:15.268031 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bv9ls_d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130/extract-content/0.log" Dec 11 09:26:15 crc kubenswrapper[4629]: I1211 09:26:15.314622 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bv9ls_d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130/extract-utilities/0.log" Dec 11 09:26:15 crc kubenswrapper[4629]: I1211 09:26:15.314758 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bv9ls_d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130/extract-content/0.log" Dec 11 09:26:15 crc kubenswrapper[4629]: I1211 09:26:15.568328 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bv9ls_d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130/extract-utilities/0.log" Dec 11 09:26:15 crc kubenswrapper[4629]: I1211 09:26:15.582061 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bv9ls_d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130/extract-content/0.log" Dec 11 09:26:15 crc kubenswrapper[4629]: I1211 09:26:15.821104 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxnk8_287e9dc5-e848-4c9c-b07e-14b244e606ef/extract-utilities/0.log" Dec 11 09:26:15 crc kubenswrapper[4629]: I1211 09:26:15.952086 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bv9ls_d1bcb2f9-52e1-4c00-a674-fb1b1b9e8130/registry-server/0.log" Dec 11 09:26:16 crc kubenswrapper[4629]: I1211 09:26:16.090073 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxnk8_287e9dc5-e848-4c9c-b07e-14b244e606ef/extract-utilities/0.log" Dec 11 09:26:16 crc kubenswrapper[4629]: I1211 09:26:16.164749 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxnk8_287e9dc5-e848-4c9c-b07e-14b244e606ef/extract-content/0.log" Dec 11 09:26:16 crc kubenswrapper[4629]: I1211 09:26:16.169627 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxnk8_287e9dc5-e848-4c9c-b07e-14b244e606ef/extract-content/0.log" Dec 11 09:26:16 crc kubenswrapper[4629]: I1211 09:26:16.403176 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxnk8_287e9dc5-e848-4c9c-b07e-14b244e606ef/extract-utilities/0.log" Dec 11 09:26:16 crc kubenswrapper[4629]: I1211 09:26:16.406122 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxnk8_287e9dc5-e848-4c9c-b07e-14b244e606ef/extract-content/0.log" Dec 11 09:26:16 crc kubenswrapper[4629]: I1211 09:26:16.784525 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4cvmx_9b48c9ef-aca6-40cf-b21c-58f804639e0e/marketplace-operator/0.log" Dec 11 09:26:16 crc kubenswrapper[4629]: I1211 09:26:16.815190 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7x2k_2f5582d1-e74c-4c04-8370-202da8ffb11f/extract-utilities/0.log" Dec 11 09:26:16 crc kubenswrapper[4629]: I1211 09:26:16.906895 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxnk8_287e9dc5-e848-4c9c-b07e-14b244e606ef/registry-server/0.log" Dec 11 09:26:17 crc kubenswrapper[4629]: I1211 09:26:17.180987 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7x2k_2f5582d1-e74c-4c04-8370-202da8ffb11f/extract-content/0.log" Dec 11 09:26:17 crc kubenswrapper[4629]: I1211 09:26:17.192565 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7x2k_2f5582d1-e74c-4c04-8370-202da8ffb11f/extract-utilities/0.log" Dec 11 09:26:17 crc kubenswrapper[4629]: I1211 09:26:17.235141 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7x2k_2f5582d1-e74c-4c04-8370-202da8ffb11f/extract-content/0.log" Dec 11 09:26:17 crc kubenswrapper[4629]: I1211 09:26:17.438469 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7x2k_2f5582d1-e74c-4c04-8370-202da8ffb11f/extract-content/0.log" Dec 11 09:26:17 crc kubenswrapper[4629]: I1211 09:26:17.460439 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7x2k_2f5582d1-e74c-4c04-8370-202da8ffb11f/extract-utilities/0.log" Dec 11 09:26:17 crc kubenswrapper[4629]: I1211 09:26:17.625760 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7x2k_2f5582d1-e74c-4c04-8370-202da8ffb11f/registry-server/0.log" Dec 11 09:26:17 crc kubenswrapper[4629]: I1211 09:26:17.699767 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jrkpq_3f21f86f-d195-44ca-9403-d5407b466aff/extract-utilities/0.log" Dec 11 09:26:17 crc kubenswrapper[4629]: I1211 09:26:17.867488 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jrkpq_3f21f86f-d195-44ca-9403-d5407b466aff/extract-content/0.log" Dec 11 09:26:17 crc kubenswrapper[4629]: I1211 09:26:17.872931 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jrkpq_3f21f86f-d195-44ca-9403-d5407b466aff/extract-utilities/0.log" Dec 11 09:26:17 crc kubenswrapper[4629]: I1211 09:26:17.925274 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jrkpq_3f21f86f-d195-44ca-9403-d5407b466aff/extract-content/0.log" Dec 11 09:26:18 crc kubenswrapper[4629]: I1211 09:26:18.147028 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jrkpq_3f21f86f-d195-44ca-9403-d5407b466aff/extract-content/0.log" Dec 11 09:26:18 crc kubenswrapper[4629]: I1211 09:26:18.152403 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jrkpq_3f21f86f-d195-44ca-9403-d5407b466aff/extract-utilities/0.log" Dec 11 09:26:18 crc kubenswrapper[4629]: I1211 09:26:18.573072 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jrkpq_3f21f86f-d195-44ca-9403-d5407b466aff/registry-server/0.log" Dec 11 09:26:20 crc kubenswrapper[4629]: I1211 09:26:20.486170 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:26:20 crc kubenswrapper[4629]: I1211 09:26:20.486237 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:26:50 crc kubenswrapper[4629]: I1211 09:26:50.486524 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:26:50 crc kubenswrapper[4629]: I1211 09:26:50.487120 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:27:20 crc kubenswrapper[4629]: I1211 09:27:20.487533 4629 patch_prober.go:28] interesting pod/machine-config-daemon-wbld9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:27:20 crc kubenswrapper[4629]: I1211 09:27:20.488135 4629 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:27:20 crc kubenswrapper[4629]: I1211 09:27:20.488198 4629 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" Dec 11 09:27:20 crc kubenswrapper[4629]: I1211 09:27:20.488987 4629 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c"} pod="openshift-machine-config-operator/machine-config-daemon-wbld9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 09:27:20 crc kubenswrapper[4629]: I1211 09:27:20.489047 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" containerName="machine-config-daemon" containerID="cri-o://224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" gracePeriod=600 Dec 11 09:27:20 crc kubenswrapper[4629]: E1211 09:27:20.621397 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:27:21 crc kubenswrapper[4629]: I1211 09:27:21.611780 4629 generic.go:334] "Generic (PLEG): container finished" podID="847b50b9-0093-404f-b0dd-63c814d8986b" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" exitCode=0 Dec 11 09:27:21 crc kubenswrapper[4629]: I1211 09:27:21.611824 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerDied","Data":"224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c"} Dec 11 09:27:21 crc kubenswrapper[4629]: I1211 09:27:21.611893 4629 scope.go:117] "RemoveContainer" containerID="d1441cf6cef81d3d1e6dde6fc7c192a9564615532b2a9036b19516be40e634b7" Dec 11 09:27:21 crc kubenswrapper[4629]: I1211 09:27:21.612566 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:27:21 crc kubenswrapper[4629]: E1211 09:27:21.612834 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:27:36 crc kubenswrapper[4629]: I1211 09:27:36.199237 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:27:36 crc kubenswrapper[4629]: E1211 09:27:36.200167 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:27:51 crc kubenswrapper[4629]: I1211 09:27:51.199556 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:27:51 crc kubenswrapper[4629]: E1211 09:27:51.200328 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:28:02 crc kubenswrapper[4629]: I1211 09:28:02.215987 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:28:02 crc kubenswrapper[4629]: E1211 09:28:02.217336 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:28:14 crc kubenswrapper[4629]: I1211 09:28:14.199171 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:28:14 crc kubenswrapper[4629]: E1211 09:28:14.200161 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:28:23 crc kubenswrapper[4629]: I1211 09:28:23.195685 4629 generic.go:334] "Generic (PLEG): container finished" podID="52f7bcbe-f979-4d26-bb6b-39aa9f251c15" containerID="1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8" exitCode=0 Dec 11 09:28:23 crc kubenswrapper[4629]: I1211 09:28:23.195787 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-65xcc/must-gather-75vsg" event={"ID":"52f7bcbe-f979-4d26-bb6b-39aa9f251c15","Type":"ContainerDied","Data":"1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8"} Dec 11 09:28:23 crc kubenswrapper[4629]: I1211 09:28:23.196677 4629 scope.go:117] "RemoveContainer" containerID="1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8" Dec 11 09:28:23 crc kubenswrapper[4629]: I1211 09:28:23.993277 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-65xcc_must-gather-75vsg_52f7bcbe-f979-4d26-bb6b-39aa9f251c15/gather/0.log" Dec 11 09:28:29 crc kubenswrapper[4629]: I1211 09:28:29.199183 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:28:29 crc kubenswrapper[4629]: E1211 09:28:29.200090 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:28:33 crc kubenswrapper[4629]: I1211 09:28:33.423167 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-65xcc/must-gather-75vsg"] Dec 11 09:28:33 crc kubenswrapper[4629]: I1211 09:28:33.424546 4629 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-65xcc/must-gather-75vsg" podUID="52f7bcbe-f979-4d26-bb6b-39aa9f251c15" containerName="copy" containerID="cri-o://e8a4a8a01ca84c5d1a064014b7f0b6af4a8522252d5be99393ebb21fb3b57b81" gracePeriod=2 Dec 11 09:28:33 crc kubenswrapper[4629]: I1211 09:28:33.447181 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-65xcc/must-gather-75vsg"] Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.033316 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-65xcc_must-gather-75vsg_52f7bcbe-f979-4d26-bb6b-39aa9f251c15/copy/0.log" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.034104 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-65xcc/must-gather-75vsg" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.169639 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/52f7bcbe-f979-4d26-bb6b-39aa9f251c15-must-gather-output\") pod \"52f7bcbe-f979-4d26-bb6b-39aa9f251c15\" (UID: \"52f7bcbe-f979-4d26-bb6b-39aa9f251c15\") " Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.170125 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5zqj\" (UniqueName: \"kubernetes.io/projected/52f7bcbe-f979-4d26-bb6b-39aa9f251c15-kube-api-access-j5zqj\") pod \"52f7bcbe-f979-4d26-bb6b-39aa9f251c15\" (UID: \"52f7bcbe-f979-4d26-bb6b-39aa9f251c15\") " Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.195932 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52f7bcbe-f979-4d26-bb6b-39aa9f251c15-kube-api-access-j5zqj" (OuterVolumeSpecName: "kube-api-access-j5zqj") pod "52f7bcbe-f979-4d26-bb6b-39aa9f251c15" (UID: "52f7bcbe-f979-4d26-bb6b-39aa9f251c15"). InnerVolumeSpecName "kube-api-access-j5zqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.275793 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5zqj\" (UniqueName: \"kubernetes.io/projected/52f7bcbe-f979-4d26-bb6b-39aa9f251c15-kube-api-access-j5zqj\") on node \"crc\" DevicePath \"\"" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.332292 4629 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-65xcc_must-gather-75vsg_52f7bcbe-f979-4d26-bb6b-39aa9f251c15/copy/0.log" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.333119 4629 generic.go:334] "Generic (PLEG): container finished" podID="52f7bcbe-f979-4d26-bb6b-39aa9f251c15" containerID="e8a4a8a01ca84c5d1a064014b7f0b6af4a8522252d5be99393ebb21fb3b57b81" exitCode=143 Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.333218 4629 scope.go:117] "RemoveContainer" containerID="e8a4a8a01ca84c5d1a064014b7f0b6af4a8522252d5be99393ebb21fb3b57b81" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.333443 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-65xcc/must-gather-75vsg" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.383402 4629 scope.go:117] "RemoveContainer" containerID="1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.507679 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52f7bcbe-f979-4d26-bb6b-39aa9f251c15-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "52f7bcbe-f979-4d26-bb6b-39aa9f251c15" (UID: "52f7bcbe-f979-4d26-bb6b-39aa9f251c15"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.512446 4629 scope.go:117] "RemoveContainer" containerID="e8a4a8a01ca84c5d1a064014b7f0b6af4a8522252d5be99393ebb21fb3b57b81" Dec 11 09:28:34 crc kubenswrapper[4629]: E1211 09:28:34.517111 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8a4a8a01ca84c5d1a064014b7f0b6af4a8522252d5be99393ebb21fb3b57b81\": container with ID starting with e8a4a8a01ca84c5d1a064014b7f0b6af4a8522252d5be99393ebb21fb3b57b81 not found: ID does not exist" containerID="e8a4a8a01ca84c5d1a064014b7f0b6af4a8522252d5be99393ebb21fb3b57b81" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.517295 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a4a8a01ca84c5d1a064014b7f0b6af4a8522252d5be99393ebb21fb3b57b81"} err="failed to get container status \"e8a4a8a01ca84c5d1a064014b7f0b6af4a8522252d5be99393ebb21fb3b57b81\": rpc error: code = NotFound desc = could not find container \"e8a4a8a01ca84c5d1a064014b7f0b6af4a8522252d5be99393ebb21fb3b57b81\": container with ID starting with e8a4a8a01ca84c5d1a064014b7f0b6af4a8522252d5be99393ebb21fb3b57b81 not found: ID does not exist" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.517418 4629 scope.go:117] "RemoveContainer" containerID="1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8" Dec 11 09:28:34 crc kubenswrapper[4629]: E1211 09:28:34.518027 4629 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8\": container with ID starting with 1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8 not found: ID does not exist" containerID="1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.518137 4629 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8"} err="failed to get container status \"1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8\": rpc error: code = NotFound desc = could not find container \"1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8\": container with ID starting with 1605eb09d03089266b1f782452ec041bbde5b1159df0f44f083af2fcf1ab3ec8 not found: ID does not exist" Dec 11 09:28:34 crc kubenswrapper[4629]: I1211 09:28:34.582579 4629 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/52f7bcbe-f979-4d26-bb6b-39aa9f251c15-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 11 09:28:36 crc kubenswrapper[4629]: I1211 09:28:36.212323 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52f7bcbe-f979-4d26-bb6b-39aa9f251c15" path="/var/lib/kubelet/pods/52f7bcbe-f979-4d26-bb6b-39aa9f251c15/volumes" Dec 11 09:28:43 crc kubenswrapper[4629]: I1211 09:28:43.199157 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:28:43 crc kubenswrapper[4629]: E1211 09:28:43.199955 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:28:57 crc kubenswrapper[4629]: I1211 09:28:57.198781 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:28:57 crc kubenswrapper[4629]: E1211 09:28:57.199671 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:29:06 crc kubenswrapper[4629]: I1211 09:29:06.585129 4629 scope.go:117] "RemoveContainer" containerID="91b2ed4bfbeffd04ea31c48c3c32e3a7db22a693f1cb82d0dec5abeacd7f825d" Dec 11 09:29:11 crc kubenswrapper[4629]: I1211 09:29:11.199067 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:29:11 crc kubenswrapper[4629]: E1211 09:29:11.199818 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:29:23 crc kubenswrapper[4629]: I1211 09:29:23.199768 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:29:23 crc kubenswrapper[4629]: E1211 09:29:23.200631 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:29:38 crc kubenswrapper[4629]: I1211 09:29:38.198888 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:29:38 crc kubenswrapper[4629]: E1211 09:29:38.199734 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:29:52 crc kubenswrapper[4629]: I1211 09:29:52.207604 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:29:52 crc kubenswrapper[4629]: E1211 09:29:52.208557 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.211784 4629 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf"] Dec 11 09:30:00 crc kubenswrapper[4629]: E1211 09:30:00.212748 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3041cd3-67f1-4459-91a2-c548576aaf76" containerName="extract-utilities" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.212760 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3041cd3-67f1-4459-91a2-c548576aaf76" containerName="extract-utilities" Dec 11 09:30:00 crc kubenswrapper[4629]: E1211 09:30:00.212782 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3041cd3-67f1-4459-91a2-c548576aaf76" containerName="extract-content" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.212788 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3041cd3-67f1-4459-91a2-c548576aaf76" containerName="extract-content" Dec 11 09:30:00 crc kubenswrapper[4629]: E1211 09:30:00.212797 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f7bcbe-f979-4d26-bb6b-39aa9f251c15" containerName="gather" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.212804 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f7bcbe-f979-4d26-bb6b-39aa9f251c15" containerName="gather" Dec 11 09:30:00 crc kubenswrapper[4629]: E1211 09:30:00.212815 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerName="extract-utilities" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.212820 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerName="extract-utilities" Dec 11 09:30:00 crc kubenswrapper[4629]: E1211 09:30:00.212837 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86c82c24-796d-4e75-8538-e75db7dd93d5" containerName="extract-content" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.212860 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="86c82c24-796d-4e75-8538-e75db7dd93d5" containerName="extract-content" Dec 11 09:30:00 crc kubenswrapper[4629]: E1211 09:30:00.212873 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86c82c24-796d-4e75-8538-e75db7dd93d5" containerName="registry-server" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.212879 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="86c82c24-796d-4e75-8538-e75db7dd93d5" containerName="registry-server" Dec 11 09:30:00 crc kubenswrapper[4629]: E1211 09:30:00.212890 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f7bcbe-f979-4d26-bb6b-39aa9f251c15" containerName="copy" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.212896 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f7bcbe-f979-4d26-bb6b-39aa9f251c15" containerName="copy" Dec 11 09:30:00 crc kubenswrapper[4629]: E1211 09:30:00.212914 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86c82c24-796d-4e75-8538-e75db7dd93d5" containerName="extract-utilities" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.212920 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="86c82c24-796d-4e75-8538-e75db7dd93d5" containerName="extract-utilities" Dec 11 09:30:00 crc kubenswrapper[4629]: E1211 09:30:00.212932 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerName="extract-content" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.212938 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerName="extract-content" Dec 11 09:30:00 crc kubenswrapper[4629]: E1211 09:30:00.212950 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerName="registry-server" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.212955 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerName="registry-server" Dec 11 09:30:00 crc kubenswrapper[4629]: E1211 09:30:00.212969 4629 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3041cd3-67f1-4459-91a2-c548576aaf76" containerName="registry-server" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.212975 4629 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3041cd3-67f1-4459-91a2-c548576aaf76" containerName="registry-server" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.213145 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3041cd3-67f1-4459-91a2-c548576aaf76" containerName="registry-server" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.213167 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="86c82c24-796d-4e75-8538-e75db7dd93d5" containerName="registry-server" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.213180 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="52f7bcbe-f979-4d26-bb6b-39aa9f251c15" containerName="gather" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.213189 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="362675ee-0b61-4101-9ea8-1b07f020bc73" containerName="registry-server" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.213199 4629 memory_manager.go:354] "RemoveStaleState removing state" podUID="52f7bcbe-f979-4d26-bb6b-39aa9f251c15" containerName="copy" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.213857 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.217204 4629 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.218719 4629 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.243296 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf"] Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.400554 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/308c4027-3e5f-499f-b5a9-dd5268fdf73d-config-volume\") pod \"collect-profiles-29424090-45bqf\" (UID: \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.400632 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjl72\" (UniqueName: \"kubernetes.io/projected/308c4027-3e5f-499f-b5a9-dd5268fdf73d-kube-api-access-xjl72\") pod \"collect-profiles-29424090-45bqf\" (UID: \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.401069 4629 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/308c4027-3e5f-499f-b5a9-dd5268fdf73d-secret-volume\") pod \"collect-profiles-29424090-45bqf\" (UID: \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.503039 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/308c4027-3e5f-499f-b5a9-dd5268fdf73d-config-volume\") pod \"collect-profiles-29424090-45bqf\" (UID: \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.503080 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjl72\" (UniqueName: \"kubernetes.io/projected/308c4027-3e5f-499f-b5a9-dd5268fdf73d-kube-api-access-xjl72\") pod \"collect-profiles-29424090-45bqf\" (UID: \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.503132 4629 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/308c4027-3e5f-499f-b5a9-dd5268fdf73d-secret-volume\") pod \"collect-profiles-29424090-45bqf\" (UID: \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.504867 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/308c4027-3e5f-499f-b5a9-dd5268fdf73d-config-volume\") pod \"collect-profiles-29424090-45bqf\" (UID: \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.518919 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/308c4027-3e5f-499f-b5a9-dd5268fdf73d-secret-volume\") pod \"collect-profiles-29424090-45bqf\" (UID: \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.531873 4629 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjl72\" (UniqueName: \"kubernetes.io/projected/308c4027-3e5f-499f-b5a9-dd5268fdf73d-kube-api-access-xjl72\") pod \"collect-profiles-29424090-45bqf\" (UID: \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:00 crc kubenswrapper[4629]: I1211 09:30:00.553334 4629 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:01 crc kubenswrapper[4629]: I1211 09:30:01.335081 4629 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf"] Dec 11 09:30:02 crc kubenswrapper[4629]: I1211 09:30:02.308334 4629 generic.go:334] "Generic (PLEG): container finished" podID="308c4027-3e5f-499f-b5a9-dd5268fdf73d" containerID="6f6aa9cba5a273759971fcb9cd01eda661e187b5761ead95d7a7b27f23c9cdb9" exitCode=0 Dec 11 09:30:02 crc kubenswrapper[4629]: I1211 09:30:02.308399 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" event={"ID":"308c4027-3e5f-499f-b5a9-dd5268fdf73d","Type":"ContainerDied","Data":"6f6aa9cba5a273759971fcb9cd01eda661e187b5761ead95d7a7b27f23c9cdb9"} Dec 11 09:30:02 crc kubenswrapper[4629]: I1211 09:30:02.308800 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" event={"ID":"308c4027-3e5f-499f-b5a9-dd5268fdf73d","Type":"ContainerStarted","Data":"b06f9fc29c204d8a4f5f59ddbca4274dcc6e83d875952c7cc437b832f5b61b9f"} Dec 11 09:30:03 crc kubenswrapper[4629]: I1211 09:30:03.677806 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:03 crc kubenswrapper[4629]: I1211 09:30:03.796566 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjl72\" (UniqueName: \"kubernetes.io/projected/308c4027-3e5f-499f-b5a9-dd5268fdf73d-kube-api-access-xjl72\") pod \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\" (UID: \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\") " Dec 11 09:30:03 crc kubenswrapper[4629]: I1211 09:30:03.796788 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/308c4027-3e5f-499f-b5a9-dd5268fdf73d-secret-volume\") pod \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\" (UID: \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\") " Dec 11 09:30:03 crc kubenswrapper[4629]: I1211 09:30:03.796974 4629 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/308c4027-3e5f-499f-b5a9-dd5268fdf73d-config-volume\") pod \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\" (UID: \"308c4027-3e5f-499f-b5a9-dd5268fdf73d\") " Dec 11 09:30:03 crc kubenswrapper[4629]: I1211 09:30:03.797918 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/308c4027-3e5f-499f-b5a9-dd5268fdf73d-config-volume" (OuterVolumeSpecName: "config-volume") pod "308c4027-3e5f-499f-b5a9-dd5268fdf73d" (UID: "308c4027-3e5f-499f-b5a9-dd5268fdf73d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:30:03 crc kubenswrapper[4629]: I1211 09:30:03.804164 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308c4027-3e5f-499f-b5a9-dd5268fdf73d-kube-api-access-xjl72" (OuterVolumeSpecName: "kube-api-access-xjl72") pod "308c4027-3e5f-499f-b5a9-dd5268fdf73d" (UID: "308c4027-3e5f-499f-b5a9-dd5268fdf73d"). InnerVolumeSpecName "kube-api-access-xjl72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:30:03 crc kubenswrapper[4629]: I1211 09:30:03.805540 4629 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308c4027-3e5f-499f-b5a9-dd5268fdf73d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "308c4027-3e5f-499f-b5a9-dd5268fdf73d" (UID: "308c4027-3e5f-499f-b5a9-dd5268fdf73d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:30:03 crc kubenswrapper[4629]: I1211 09:30:03.899242 4629 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/308c4027-3e5f-499f-b5a9-dd5268fdf73d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 09:30:03 crc kubenswrapper[4629]: I1211 09:30:03.899290 4629 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjl72\" (UniqueName: \"kubernetes.io/projected/308c4027-3e5f-499f-b5a9-dd5268fdf73d-kube-api-access-xjl72\") on node \"crc\" DevicePath \"\"" Dec 11 09:30:03 crc kubenswrapper[4629]: I1211 09:30:03.899308 4629 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/308c4027-3e5f-499f-b5a9-dd5268fdf73d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 09:30:04 crc kubenswrapper[4629]: I1211 09:30:04.328633 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" event={"ID":"308c4027-3e5f-499f-b5a9-dd5268fdf73d","Type":"ContainerDied","Data":"b06f9fc29c204d8a4f5f59ddbca4274dcc6e83d875952c7cc437b832f5b61b9f"} Dec 11 09:30:04 crc kubenswrapper[4629]: I1211 09:30:04.328684 4629 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b06f9fc29c204d8a4f5f59ddbca4274dcc6e83d875952c7cc437b832f5b61b9f" Dec 11 09:30:04 crc kubenswrapper[4629]: I1211 09:30:04.328981 4629 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424090-45bqf" Dec 11 09:30:04 crc kubenswrapper[4629]: I1211 09:30:04.786030 4629 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q"] Dec 11 09:30:04 crc kubenswrapper[4629]: I1211 09:30:04.800302 4629 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424045-lkk6q"] Dec 11 09:30:05 crc kubenswrapper[4629]: I1211 09:30:05.199100 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:30:05 crc kubenswrapper[4629]: E1211 09:30:05.199346 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:30:06 crc kubenswrapper[4629]: I1211 09:30:06.214504 4629 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c13206c5-13b9-4bf3-bdfd-7dd30d2ce643" path="/var/lib/kubelet/pods/c13206c5-13b9-4bf3-bdfd-7dd30d2ce643/volumes" Dec 11 09:30:06 crc kubenswrapper[4629]: I1211 09:30:06.659135 4629 scope.go:117] "RemoveContainer" containerID="eafab88ed89af90c86985f125c9b33a303b00063ec91c45f41c68421b6e95218" Dec 11 09:30:17 crc kubenswrapper[4629]: I1211 09:30:17.199425 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:30:17 crc kubenswrapper[4629]: E1211 09:30:17.200351 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:30:31 crc kubenswrapper[4629]: I1211 09:30:31.198967 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:30:31 crc kubenswrapper[4629]: E1211 09:30:31.199711 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:30:45 crc kubenswrapper[4629]: I1211 09:30:45.199255 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:30:45 crc kubenswrapper[4629]: E1211 09:30:45.200206 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:31:00 crc kubenswrapper[4629]: I1211 09:31:00.199937 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:31:00 crc kubenswrapper[4629]: E1211 09:31:00.201115 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:31:06 crc kubenswrapper[4629]: I1211 09:31:06.713767 4629 scope.go:117] "RemoveContainer" containerID="fc065bb6c85f2ecb4f591d32170016cc1598ace86f844c50912b07cc45bc9609" Dec 11 09:31:06 crc kubenswrapper[4629]: I1211 09:31:06.741726 4629 scope.go:117] "RemoveContainer" containerID="e5855caf15e7a4089d20234a8764b0d6f5c24288f84571e37b9a3560ade421c5" Dec 11 09:31:06 crc kubenswrapper[4629]: I1211 09:31:06.800685 4629 scope.go:117] "RemoveContainer" containerID="072acea5f332625a159aca198441980ff25f5eb7070a86bfcd0fd32112864d7d" Dec 11 09:31:13 crc kubenswrapper[4629]: I1211 09:31:13.200627 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:31:13 crc kubenswrapper[4629]: E1211 09:31:13.201561 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:31:26 crc kubenswrapper[4629]: I1211 09:31:26.199221 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:31:26 crc kubenswrapper[4629]: E1211 09:31:26.200139 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:31:41 crc kubenswrapper[4629]: I1211 09:31:41.199715 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:31:41 crc kubenswrapper[4629]: E1211 09:31:41.200566 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:31:53 crc kubenswrapper[4629]: I1211 09:31:53.198684 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:31:53 crc kubenswrapper[4629]: E1211 09:31:53.199467 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:32:04 crc kubenswrapper[4629]: I1211 09:32:04.199747 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:32:04 crc kubenswrapper[4629]: E1211 09:32:04.200703 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:32:16 crc kubenswrapper[4629]: I1211 09:32:16.199247 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:32:16 crc kubenswrapper[4629]: E1211 09:32:16.200159 4629 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wbld9_openshift-machine-config-operator(847b50b9-0093-404f-b0dd-63c814d8986b)\"" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" podUID="847b50b9-0093-404f-b0dd-63c814d8986b" Dec 11 09:32:30 crc kubenswrapper[4629]: I1211 09:32:30.200996 4629 scope.go:117] "RemoveContainer" containerID="224367886b509cb8c313ba170ec9a3f0d5e41b2154eea1871195c5f0bf423c6c" Dec 11 09:32:30 crc kubenswrapper[4629]: I1211 09:32:30.727763 4629 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wbld9" event={"ID":"847b50b9-0093-404f-b0dd-63c814d8986b","Type":"ContainerStarted","Data":"6c30709547bf74383cd0e10a3694d3d15dfae12062fe4f9db63247d6caac55c0"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515116510030024435 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015116510031017353 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015116500554016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015116500555015460 5ustar corecore